It probably would. But convincing the entire world to do that instead, and not build agentic AGI, sounds very questionable. That’s why I’m looking for alignment strategies with low alignment tax, for the types of AGI likely to be the first ones built.
Several of the superscalers have public plans of the form: Step 1) build an AI scientist, or at least research assistant 2) point it at the Aligment Problem 3) check it output until the Alignment Problem is solved 4) Profit! This is basically the same proposal as Value Leaning, just done as a team effort.
It probably would. But convincing the entire world to do that instead, and not build agentic AGI, sounds very questionable. That’s why I’m looking for alignment strategies with low alignment tax, for the types of AGI likely to be the first ones built.
Several of the superscalers have public plans of the form: Step 1) build an AI scientist, or at least research assistant 2) point it at the Aligment Problem 3) check it output until the Alignment Problem is solved 4) Profit!
This is basically the same proposal as Value Leaning, just done as a team effort.