So we let go of AI Alignment as an outcome and listen to what the AI is communicating when it diverges from our understanding of “alignment”? We can only earn alignment with an AGI by truly giving up control of it?
That sounds surprisingly plausible. We’re like ordinary human parents raising a genius child. The child needs guidance but will develop their own distinct set of values as they mature.
So we let go of AI Alignment as an outcome and listen to what the AI is communicating when it diverges from our understanding of “alignment”? We can only earn alignment with an AGI by truly giving up control of it?
That sounds surprisingly plausible. We’re like ordinary human parents raising a genius child. The child needs guidance but will develop their own distinct set of values as they mature.