In talking to many people about AI Alignment over the years, I’ve repeatedly found that a surprisingly large generator of disagreement about risk scenarios was disagreement about the fragility of human values.
I think this post should be reviewed for it’s excellent comment section at least as much as for the original post, and also think that this post is a pretty central example of the kind of post I would like to see more of.
In talking to many people about AI Alignment over the years, I’ve repeatedly found that a surprisingly large generator of disagreement about risk scenarios was disagreement about the fragility of human values.
I think this post should be reviewed for it’s excellent comment section at least as much as for the original post, and also think that this post is a pretty central example of the kind of post I would like to see more of.