I’ll be happy if AI gives people time/space/safety to figure out what they want while taking actions in the world that preserve option value.
The kind of AI alignment solution we’re working on isn’t a substitute for people deciding how they want to reflect and develop and decide what they value. The idea is that if AI is going to be part of that process, then the timing and nature of AI involvement should be decided by people rather than by “we need to deploy this AI now in order to remain competitive and accept whatever affects that has on our values.”
You could imagine AI solutions that try to replace the normal process of moral deliberation and reconciliation (rather than simply being a tool to help it), but I’ve never seen a proposal along those lines that didn’t seem really bad to me.
I’ll be happy if AI gives people time/space/safety to figure out what they want while taking actions in the world that preserve option value.
The kind of AI alignment solution we’re working on isn’t a substitute for people deciding how they want to reflect and develop and decide what they value. The idea is that if AI is going to be part of that process, then the timing and nature of AI involvement should be decided by people rather than by “we need to deploy this AI now in order to remain competitive and accept whatever affects that has on our values.”
You could imagine AI solutions that try to replace the normal process of moral deliberation and reconciliation (rather than simply being a tool to help it), but I’ve never seen a proposal along those lines that didn’t seem really bad to me.