I’ve used the concepts in this post a lot when discussing various things related to AI Alignment. I think asking “how robust is this AI design to various ways of scaling up?” has become one of my go-to hammers for evaluating a lot of AI Alignment proposals, and I’ve gotten a lot of mileage out of that.
I’ve used the concepts in this post a lot when discussing various things related to AI Alignment. I think asking “how robust is this AI design to various ways of scaling up?” has become one of my go-to hammers for evaluating a lot of AI Alignment proposals, and I’ve gotten a lot of mileage out of that.