Is There a Power Play Overhang?

Link post

This post is about risks in the development of increasingly capable AI, in particular the risk of losing control to AI and extinction risk. I’ll suggest that a key question is, “When do we need to take this kind of risk seriously?”

We’ll look at the issue of ‘agency overhang’, which suggests that adding agent-like abilities to AI could result in a sudden and surprising increase in these kinds of risks.

I’ll draw on intuitions about humans taking administrative and political control (with reference to the ‘Dictator Book Club’) and rephrase agency overhang as ‘power play overhang’.

I’ll finish by suggesting that a lot of people may be making a subtle but important mistake in imagining just one fairly specific path to dangerous AI.

No comments.