Another factor for the safest type of AGI is one that can practically be built soon.
The perfect is the enemy of the good. A perfectly safe system that will be deployed five years after the first self-improving AGI is probably useless.
Of course the safest path is to never build an agentic AGI. But that seems unlikely.
This criteria is another argument for language model agents. I’ve outlined their list of safety advantages here.
Of course, we don’t know if language model agents will achieve full AGI.
Another path to AGI that seems both achievable and alignable is loosely brainlike AGI, along the lines of LeCun’s proposed H-JEPA. Steve Byrnes’ “plan for mediocre alignment” seems extensible to become quite a good plan for this type of AGI.
Another factor for the safest type of AGI is one that can practically be built soon.
The perfect is the enemy of the good. A perfectly safe system that will be deployed five years after the first self-improving AGI is probably useless.
Of course the safest path is to never build an agentic AGI. But that seems unlikely.
This criteria is another argument for language model agents. I’ve outlined their list of safety advantages here.
Of course, we don’t know if language model agents will achieve full AGI.
Another path to AGI that seems both achievable and alignable is loosely brainlike AGI, along the lines of LeCun’s proposed H-JEPA. Steve Byrnes’ “plan for mediocre alignment” seems extensible to become quite a good plan for this type of AGI.