“it also seems quite likely (though not certain) that Eliezer was wrong about how hard Aimability/Control actually is”
This seems significant. Could you elaborate? How hard do you think amiability/control is? Why do you think this is true? Who else seems to think the same?
Or see almost every post labeled Aligned AI proposals (including some from me). Most of which are based on specific concrete implementations of AI, such as LLMs, having possibly-useful alignment aimability properties that the abstract worst-case assumptions about the outcome of Reinforcement Learning that LW/MIDI were thinking about a decade ago don’t.
“it also seems quite likely (though not certain) that Eliezer was wrong about how hard Aimability/Control actually is”
This seems significant. Could you elaborate? How hard do you think amiability/control is? Why do you think this is true? Who else seems to think the same?
See the AI optimists site
https://optimists.ai/2023/11/28/ai-is-easy-to-control/
Or see almost every post labeled Aligned AI proposals (including some from me). Most of which are based on specific concrete implementations of AI, such as LLMs, having possibly-useful alignment aimability properties that the abstract worst-case assumptions about the outcome of Reinforcement Learning that LW/MIDI were thinking about a decade ago don’t.