I don’t, because as far as I understand it there is no principal/agent mechanism at work in a Stag Hunt. I can see I was powerfully vague though, so thank you for pointing that out via the question.
I was comparing Stag Hunt to the Prisoner’s Dilemma, and the argument is this:
Prisoner’s Dilemma is one agent reasoning about another agent. This is simple, so there will be many papers on it.
Stag Hunt is multiple agents reasoning about multiple agents. This is less simple, so there will be fewer papers, corresponding to the difference in difficulty.
I expect the same to also apply to the transition from one principal and one agent to multiple principles and multiple agents.
Returning to the sufficiency claim: I think I weigh the “Alignment framings from MIRIs early years” arguments more heavily than Andrew does; I estimate a mild over-commitment to the simplest-case-first-norm of approximately the same strength as the community’s earlier over-commitment to modest epistemology would be sufficient to explain the collective “ugh” response. It’s worth noting that the LessWrong sector is the only one referenced that has much in the way of laypeople—which is to say people like me—in it. I suspect that our presence in the community biases it more strongly towards simpler procedures, which leads me to put more weight on the over-commitment explanation.
That being said, my yeoman-community-member impressions of the anti-politics bias largely agree with Andrew’s, even though I only read this website, and some of the high-level discussion of papers and research agenda posts from MIRI/Open AI/Deepmind/etc. My gut feeling says there should be a way to make multi/multi AI dynamics palatable for us despite this. For example, consider the popularity of posts surrounding Voting Theory, which are all explicitly political. Multi/multi dynamics are surely less political than that, I reason.
I don’t, because as far as I understand it there is no principal/agent mechanism at work in a Stag Hunt. I can see I was powerfully vague though, so thank you for pointing that out via the question.
I was comparing Stag Hunt to the Prisoner’s Dilemma, and the argument is this:
Prisoner’s Dilemma is one agent reasoning about another agent. This is simple, so there will be many papers on it.
Stag Hunt is multiple agents reasoning about multiple agents. This is less simple, so there will be fewer papers, corresponding to the difference in difficulty.
I expect the same to also apply to the transition from one principal and one agent to multiple principles and multiple agents.
Returning to the sufficiency claim: I think I weigh the “Alignment framings from MIRIs early years” arguments more heavily than Andrew does; I estimate a mild over-commitment to the simplest-case-first-norm of approximately the same strength as the community’s earlier over-commitment to modest epistemology would be sufficient to explain the collective “ugh” response. It’s worth noting that the LessWrong sector is the only one referenced that has much in the way of laypeople—which is to say people like me—in it. I suspect that our presence in the community biases it more strongly towards simpler procedures, which leads me to put more weight on the over-commitment explanation.
That being said, my yeoman-community-member impressions of the anti-politics bias largely agree with Andrew’s, even though I only read this website, and some of the high-level discussion of papers and research agenda posts from MIRI/Open AI/Deepmind/etc. My gut feeling says there should be a way to make multi/multi AI dynamics palatable for us despite this. For example, consider the popularity of posts surrounding Voting Theory, which are all explicitly political. Multi/multi dynamics are surely less political than that, I reason.