Doc Xardoc reports back on the Chinese alignment overview paper that it mostly treats alignment as an incidental engineering problem, at about a 2.5 on a 1-10 scale with Yudkowsky being 10
I’m pretty sure Yudkowsky is at around an 8.5 actually (I think he thinks it’s not impossible in principle for ML like systems but maybe it is). 10 would be impossible in principle.
I’m pretty sure Yudkowsky is at around an 8.5 actually (I think he thinks it’s not impossible in principle for ML like systems but maybe it is). 10 would be impossible in principle.