Depends on if you care about outcome or morality. Already now we use terms like punishment, reward and adverserial competition between neural networks.
If the entity (be it a corporation or AGI-driven corporation) learns from the punishment (or its peers learn by observation) it doesn't matter if the entity is sentient, the outcome is less of the unwanted behaviour. I didn't mean anything more complicated than that.
The corporation cannot learn. Why anthropomorphise when you can actually think about the real problem: how to incentivise the people who own and direct companies to behave well.