Conditional on the AI never doing something like: manipulating/deceiving[1] the humans such that the humans think the AI is aligned, such that the AI can later do things the humans don’t like, then I am much more optimistic about the whole situation.
Conditional on the AI never doing something like: manipulating/deceiving[1] the humans such that the humans think the AI is aligned, such that the AI can later do things the humans don’t like, then I am much more optimistic about the whole situation.
The AI could be on some level not “aware” that it was deceiving the humans, a la Deep Deceptiveness.