Wow, there’s a lot of ground to cover. For everyone who hasn’t read Eliezer’s previous writings, he talks about something very similar in Creating Friendly Artificial Intelligence, all the way back in 2001 (link = http://www.singinst.org/upload/CFAI/design/structure/external.html). With reference to Andy Wood’s comment:
“What claim could any person or group have to landing closer to the one-place function?”
Next obvious question: For purposes of Friendly AI, and for correcting mistaken intuitions, how do we approximate the rightness function? How do we determine whether A(x) or B(x) is a closer approximation to Right(x)?
Next obvious answer: The rightness function can be computed by computing humanity’s Coherent Extrapolated Volition, written about by Eliezer in 2004 (http://www.singinst.org/upload/CEV.html). The closer a given algorithm comes to humanity’s CEV, the closer it should come to Right(x).
Note: I did not think of CFAI when I read Eliezer’s previous post, although I did think of CEV as a candidate for morality’s content. CFAI refers to the supergoals of agents in general, while all the previous posts referred to a tangle of stuff surrounding classic philosophical ideas of morality, so I didn’t connect the dots.
Wow, there’s a lot of ground to cover. For everyone who hasn’t read Eliezer’s previous writings, he talks about something very similar in Creating Friendly Artificial Intelligence, all the way back in 2001 (link = http://www.singinst.org/upload/CFAI/design/structure/external.html). With reference to Andy Wood’s comment:
“What claim could any person or group have to landing closer to the one-place function?”
Next obvious question: For purposes of Friendly AI, and for correcting mistaken intuitions, how do we approximate the rightness function? How do we determine whether A(x) or B(x) is a closer approximation to Right(x)?
Next obvious answer: The rightness function can be computed by computing humanity’s Coherent Extrapolated Volition, written about by Eliezer in 2004 (http://www.singinst.org/upload/CEV.html). The closer a given algorithm comes to humanity’s CEV, the closer it should come to Right(x).
Note: I did not think of CFAI when I read Eliezer’s previous post, although I did think of CEV as a candidate for morality’s content. CFAI refers to the supergoals of agents in general, while all the previous posts referred to a tangle of stuff surrounding classic philosophical ideas of morality, so I didn’t connect the dots.