The only tone that bothers me about SR is the whole “Let’s be hippies and work everything out and it’ll all be ok” thing. “Free Lunch” in particular.
Are we both thinking of the book where vg gnxrf n qrhf rk znpuvan gb cerirag uhznavgl sebz qrfgeblvat vgfrys? Gur obbx va juvpu ng yrnfg bar punenpgre’f rkgencbyngrq ibyvgvba jbhyq cebonoyl qrfgebl uhznavgl, cnvashyyl?
Now the AI does seem absurd. I’m tempted to give SR a pass on that one because he had the characters talk about science fiction so much, they almost break the fourth wall to explain his motives. But the same author went on a rant elsewhere about the dangers of Star Trek science fantasy. His apparent exception for Callahan’s seems a little forced.
Are we both thinking of the book where vg gnxrf n qrhf rk znpuvan gb cerirag uhznavgl sebz qrfgeblvat vgfrys? Gur obbx va juvpu ng yrnfg bar punenpgre’f rkgencbyngrq ibyvgvba jbhyq cebonoyl qrfgebl uhznavgl, cnvashyyl?
When and which character? I’m not sure where you’re getting that.
Now the AI does seem absurd.
Well, it’s a pretty common error to think that with enough intelligence, an AI (or person) will be ethical and friendly. Eliezer himself made that mistake back in 2000 before he realized that intelligence is optimizing the world towards goals and those goals can be arbitrary. Spider was right that an AI would probably not have human emotions like greed or revenge, but he missed the idea that we’re made of atoms that the AI could use for something else.
Are we both thinking of the book where vg gnxrf n qrhf rk znpuvan gb cerirag uhznavgl sebz qrfgeblvat vgfrys? Gur obbx va juvpu ng yrnfg bar punenpgre’f rkgencbyngrq ibyvgvba jbhyq cebonoyl qrfgebl uhznavgl, cnvashyyl?
Now the AI does seem absurd. I’m tempted to give SR a pass on that one because he had the characters talk about science fiction so much, they almost break the fourth wall to explain his motives. But the same author went on a rant elsewhere about the dangers of Star Trek science fantasy. His apparent exception for Callahan’s seems a little forced.
When and which character? I’m not sure where you’re getting that.
Well, it’s a pretty common error to think that with enough intelligence, an AI (or person) will be ethical and friendly. Eliezer himself made that mistake back in 2000 before he realized that intelligence is optimizing the world towards goals and those goals can be arbitrary. Spider was right that an AI would probably not have human emotions like greed or revenge, but he missed the idea that we’re made of atoms that the AI could use for something else.
In the book you’re talking about, what do we learn in the big reveal? What happens immediately after the big reveal? Do we both mean this book?
Yes, that book. By big reveal, do you mean gung gur vagehqref ner gvzr geniryref? Please elaborate.
Jura Ubezng gur gvzr geniryre rkcynvaf uvf zbgvirf, ur rkcyvpvgyl fnlf gur uhzna enpr vf “qbbzrq” va uvf gvzr. Gurl pna’g ercebqhpr cebcreyl, naq gurl qba’g frrz gb unir nal cebfcrpgf sbe vzzbegnyvgl. Gurl ubcr gvzr geniry jbexf va rknpgyl gur evtug jnl gb yrg gurz punatr uvfgbel sbe gur orggre, orpnhfr jung qb gurl unir gb ybfr? V pnyyrq guvf n qrhf rk znpuvan (nffhzvat vg jbexf).
Nsgre Ubezng rkcynvaf rirelguvat, gur onq thl’f Qentba be ungpurg-zna erirnyf gung ur urneq vg nf jryy naq cynaf gb xvyy gurz. Ur frrzf snveyl vagryyvtrag, pregnvayl fznegre guna uvf rzcyblre. Ohg ur oryvrirf ur cersref n jbeyq jurer ur trgf gb xrrc gur wbo ur ybirf, naq yngre nyy uhznaf qvr.