Layperson question here; I appreciate any answers/comments that might help me understand this topic better.
When thinking about DALL·E 2 and Google’s PaLM this week I wondered if in the future we could hypothetically have AI systems like this...
Input: “an award-winning 2-hour sci-fi drama feature film about blah blah blah”
Output: an incredible film that people living in the 1970s would assume was made by future humans, not future AI, and deserves to win Best Picture because it made them laugh and cry and was amazing
...all without having to solve the alignment problem because the system that produces the film isn’t actually intelligent in the ways that matter to AI risk.
That is, can we get crazy impressive outputs from an AI system without that AI system posing an existential risk or being aligned?
If so, what feature distinguishes AI systems that do pose existential risks and need to be aligned from those that don’t?
If not, what necessary aspect of the AI system that produces the output above is it that makes it so the system now poses an existential risk and needs to be aligned?
Yes. Deep Blue was impressive in 1997.
Generality + intelligence. Deep Blue was domain-specific. Your laptop computer is perfectly general but has little intelligence.
None. Neither pose an existential risk.
The relevant tradeoff to consider is the cost of prediction and the cost of influence. As long as the cost of predicting an “impressive output” is much lower than the cost of influencing the world such that an easy-to-generate output is considered impressive, then it’s possible to generate the impressive output without risking misalignment by bounding optimization power at lower than the power required to influence the world.
So you can expect an impressive AI that predicts the weather but isn’t allowed to, e.g., participate in prediction markets on the weather nor charter flights to seed clouds to cause rain, without needing to worry about alignment. But don’t expect alignment-irrelevance from a bot aimed at writing persuasive philosophical essays, nor an AI aimed at predicting the behavior of the stock market conditional on the trades it tells you to make, nor an AI aimed at predicting the best time to show you an ad for the AI’s highest-paying company.