So while an SAI might acquire resources, it’s not clear an aligned SAI would keep humans in control, for different values of ‘aligned’ and ‘in control’.
I was referring to Paul’s own approach to AI alignment, which does aim to keep humans in control. See this post where he mentions this, and perhaps this recent overview of Paul’s approach if you’re not familiar with it.
I was referring to Paul’s own approach to AI alignment, which does aim to keep humans in control. See this post where he mentions this, and perhaps this recent overview of Paul’s approach if you’re not familiar with it.
Thanks for clarifying. I didn’t know you were specifically referring to Paul’s approach. I’ve got familiarize myself with it more.