I have, completely seriously, been thinking that uploading a human who is aligned with human values (Eliezer Yudkowsky?) might be the safest option for creating FAI. Though I didn’t think of it is a serious possibility technically speaking, in the kind of timeframe we have available (and still not sure it is. how do I evaluate that?).
I have also been thinking about furthering my understanding of Friendliness. Human ethics. What would we want FAI to do? (the part of it that is possible for a human to understand, anyway) since that seems important both for AI alignment and for human alignment.
An alternative idea I’ve had is if it’s possible for a human to reason well about human ethics and to answer difficult moral questions about the kind of decisions FAI would need to make, we explore the mental processes of such humans (who are good moral reasoners) and “scale them up” to build FAI. Maybe literally (brain-like AGI?).
I don’t know about human values, but Yudkovsky is certainly not aligned with my values on at least a couple significant points.
Still beats either Clippy or Xi Jinping of course.
I have, completely seriously, been thinking that uploading a human who is aligned with human values (Eliezer Yudkowsky?) might be the safest option for creating FAI. Though I didn’t think of it is a serious possibility technically speaking, in the kind of timeframe we have available (and still not sure it is. how do I evaluate that?).
I have also been thinking about furthering my understanding of Friendliness. Human ethics. What would we want FAI to do? (the part of it that is possible for a human to understand, anyway) since that seems important both for AI alignment and for human alignment.
An alternative idea I’ve had is if it’s possible for a human to reason well about human ethics and to answer difficult moral questions about the kind of decisions FAI would need to make, we explore the mental processes of such humans (who are good moral reasoners) and “scale them up” to build FAI.
Maybe literally (brain-like AGI?).
I don’t know about human values, but Yudkovsky is certainly not aligned with my values on at least a couple significant points. Still beats either Clippy or Xi Jinping of course.