There’s a lot of stuff I agree with in your post, but one thing I disagree with is point 3. See Where do you get your capabilities from?, especially the bounded breakdown of the orthogonality thesis part at the end.
Not that I think this makes GPT models fully safe, but I think its unsafety will look a lot more like the unsafety of humans, plus some changes in the price of things. (Which can make a huge difference.)
There’s a lot of stuff I agree with in your post, but one thing I disagree with is point 3. See Where do you get your capabilities from?, especially the bounded breakdown of the orthogonality thesis part at the end.
Not that I think this makes GPT models fully safe, but I think its unsafety will look a lot more like the unsafety of humans, plus some changes in the price of things. (Which can make a huge difference.)