I wouldn’t be so quick to discard the idea of the AI persuading us that things are pretty nice the way they are.
Suppose the AI we build (AI1) finds itself insufficiently intelligent to persuade us. It decides to build a more powerful AI (AI2) to give it advice. AI2 wakes up and modifies AI1 into being perfectly satisfied with the way things are. Then, mission accomplished, they both shut down and leave humanity unchanged.
I think what went wrong here is that this formulation of utilitarianism isn’t reflectively consistent.
There are probably strong limits to the persuadability of human beings, so it wouldn’t be a disaster.
If there are, then the AI would modify us physically instead.
Suppose the AI we build (AI1) finds itself insufficiently intelligent to persuade us. It decides to build a more powerful AI (AI2) to give it advice. AI2 wakes up and modifies AI1 into being perfectly satisfied with the way things are. Then, mission accomplished, they both shut down and leave humanity unchanged.
I think what went wrong here is that this formulation of utilitarianism isn’t reflectively consistent.
If there are, then the AI would modify us physically instead.