Hello. I am Ella Markianos. I am an undergraduate student trying to understand alignment. If you like or dislike any of my posts please email me at apollonianblues@gmail.com; I’ve always wanted a pen pal.
apollonianblues
Karma: 33
I have LOL thanks tho
My assumption is that it would do this to prevent other people from making superintelligences that are unaligned. At least Eliezer thinks you need to do this (see bullet point 6 in this post), and I think it generally comes up in conversations people have about pivotal acts. Some people think if you think of an alignment solution that’s good and easy to implement, everyone building AGI will use it, and so you won’t have to prevent other people from building unaligned AGI, but this seems unrealistic and risky to me.
TBH my naive thought is that if John’s project succeeds it’ll solve most of what I think of as the hard part of alignment, and so it seems like one of the more promising approaches to me, but in my model of the world it seems quite unlikely that there are natural abstractions in the way that John seems to think there are.