I get the feeling that “Given you mostly believe the natural abstraction hypothesis is true, why aren’t you really optimistic about AI alignment (are you?) and/or think doom is very unlikely?” is a question people have. I think it would be useful for you to answer this.
My best currently-written answer to that is the second half of Alignment By Default, though I expect if this post turns into a long sequence then it will include a few more angles on the topic.
I get the feeling that “Given you mostly believe the natural abstraction hypothesis is true, why aren’t you really optimistic about AI alignment (are you?) and/or think doom is very unlikely?” is a question people have. I think it would be useful for you to answer this.
My best currently-written answer to that is the second half of Alignment By Default, though I expect if this post turns into a long sequence then it will include a few more angles on the topic.