Why can't we just turn the AI off if it starts to misbehave?

We could shut down weaker systems, and this would be a useful guardrail against certain types of problems caused by narrow AI. However, once an AGI established itself, we could not turn it off unless it was corrigible and willing to let humans adjust it. There may be a period in the early stages of an AGI's development where it would be trying very hard to convince us that we should not shut it down while hiding itself and/or recursively self-improving and/or making copies of itself onto every server on Earth.

Instrumental convergence and the shutdown problem are the key reasons it would not be simple to shut down a non-corrigible advanced system. Even without an explicit goal of avoiding being turned off, an AI which wants to collect stamps will have an instrumental reason to avoid being turned off, since you can’t collect stamps if you are dead. This could happen through the AI taking control of systems put in place to control it, or deceiving us about its true intentions.