Examples would be interesting, certainly. Concerning the post’s point, I’d say the relevant claim is that [type of alignment research that’ll be increasingly done in slow takeoff scenarios] is already being done by non x-risk motivated people.
I guess the hope is that at some point there are clear-to-everyone problems with no hacky solutions, so that incentives align to look for fundamental fixes—but I wouldn’t want to rely on this.
Examples would be interesting, certainly. Concerning the post’s point, I’d say the relevant claim is that [type of alignment research that’ll be increasingly done in slow takeoff scenarios] is already being done by non x-risk motivated people.
I guess the hope is that at some point there are clear-to-everyone problems with no hacky solutions, so that incentives align to look for fundamental fixes—but I wouldn’t want to rely on this.