Why has EV made many moves in the direction of decentralizing EA, rather than in the direction of centralizing it?
Power within EA is currently highly centralized. It seems very likely that the correct amount of centralization is less than the current amount.
Why, as an organization aiming to ensure the health of a community that is majority male and includes many people of color, does the CEA Community Health team consist of seven white women, no men, and no people of color?
This sounds like a rhetorical question. The non-rhetorical answer is that women are much more likely than men to join a Community Health team, for approximately the same reason that most companies’ HR teams are mostly women; and nobody has bothered to counteract this.
Has anyone considered possible perverse incentives that the aforementioned CEA Community Health team may experience, in that they may have incentives to exaggerate problems in the community to justify their own existence?
I had never considered that but I don’t think it’s a strong incentive. It doesn’t look like the Community Health team is doing this. If anything, I think they’re incentivized to give themselves less work, not more.
Why do very few EA organizations do large mainstream fundraising campaigns outside the EA community, when the vast majority of outside charities do?
That’s not correct. Lots of EA orgs fundraise outside of the EA community.
Why have so few people, both within EA and within popular discourse more broadly, drawn parallels between the “TESCREAL” conspiracy theory and antisemitic conspiracy theories?
Because guilt-by-association is a very weak form of argument. (And it’s not even obvious to me that there are relevant parallels there.) And FWIW I don’t respond to the sorts of people who use the word “TESCREAL” because I don’t think they’re worth taking seriously.
Why do university EA groups appear, at least upon initial examination, to focus so much on recruiting, to the exclusion of training students and connecting them with interested people?
University groups do do those other things. But they do those things internally so you don’t notice. Recruiting is the only thing they do externally, so that’s what you notice.
Why aren’t there more organizations within EA that are trying to be extremely hardcore and totalizing, to the level of religious orders, the Navy SEALs, the Manhattan Project, or even a really intense start-up?
Some orgs did that and it generally didn’t go well (eg Leverage Research). I think most people believe that totalizing jobs are bad for mental health and create bad epistemics and it’s not worth it.
When EAs talk about the “unilateralist’s curse,” why don’t they qualify those claims with the fact that Arkhipov and Petrov were unilateralists who likely saved the world from nuclear war?
Those are not examples of the unilateralist’s curse. I don’t want to explain it in this short comment but I would suggest re-reading some materials that explain the unilateralist’s curse, e.g. the original paper.
Why hasn’t AI safety as a field made an active effort to build large hubs outside the Bay, rather than the current state of affairs in which outside groups basically just function as recruiting channels to get people to move to the Bay?
Because doing so would be a lot of work, which would take time away from doing other important things. I think people agree that having a second hub would be good, but not good enough to justify the effort.
Not necessarily guilt-by-association, but maybe rather pointing out that the two arguments/conspiracy theories share a similar flawed structure, so if you discredit one, you should discredit the other.
Still, I’m also unsure how much structure they share, and even if they did, I don’t think this would be discursively effective because I don’t think most people care that much about (that kind of) consistency (happy to be updated in the direction of most people caring about it).
Thanks for giving some answers here to these questions; it was really helpful to have them laid out like this.
1. In hindsight, I was probably talking more about moves towards decentralization of leadership, rather than decentralization of funding. I agree that greater decentralization of funding is a good thing, but it seems to me like, within the organizations funded by a given funder, decentralization of leadership is likely useless (if leadership decisions are still being made by informal networks between orgs rather than formal ones), or it may lead to a lack of clarity and direction.
3. I understand the dynamics that may cause the overrepresentation of women. However, that still doesn’t completely explain why there is an overrepresentation of white women, even when compared to racial demographics within EA at large. Additionally, this also doesn’t explain why the overrepresentation of women here isn’t seen as a problem on CEA’s part, if even just from an optics perspective.
4. Makes sense, but I’m still concerned that, say, if CEA had an anti-Stalinism team, they’d be reluctant to ever say “Stalinism isn’t a problem in EA.”
5. Again, this was a question that was badly worded on my end. I was referring more specifically to organizations within AI safety, more than EA at large. I know that AMF, GiveDirectly, The Humane League, etc. fundraise outside EA.
6. I was asking a descriptive question here, not a normative one. Guilt by association, even if weak, is a very commonly used form of argument, and so I would expect it to be in used in this case.
7. That makes sense. That was one of my hypotheses (hence my phrase “at least upon initial examination”), and I guess in hindsight it’s probably the best one.
10. Starting an AI capabilities company that does AI safety as a side project generally hasn’t gone well, and yet people keep doing it. The fact that something hasn’t gone well in the past doesn’t seem to me to be a sufficient explanation for why people don’t keep doing it, especially because it largely seems like Leverage failed for Leverage-specific reasons (i.e. too much engagement with woo). Additionally, your argument here seems to prove too much; the Manhattan Project was a large scientific project operating under an intense structure, and yet it was able to maintain good epistemics (i.e. not fixating too hard on designs that wouldn’t work) under those conditions. Same with a lot of really intense start-ups.
11. They may not be examples of the unilateralist’s curse in the original sense, but the term seems to have been expanded well past its original meaning, and they’re examples of that expanded meaning.
12. It seems to me like this is work of a different type than technical alignment work, and could likely be accomplished by hiring different people than the people already working on technical alignment, so it’s not directly trading off against that.
I was asking a descriptive question here, not a normative one. Guilt by association, even if weak, is a very commonly used form of argument, and so I would expect it to be in used in this case.
I intended my answer to be descriptive. EAs generally avoid making weak arguments (or at least I like to think we do).
I will attempt to answer a few of these.
Power within EA is currently highly centralized. It seems very likely that the correct amount of centralization is less than the current amount.
This sounds like a rhetorical question. The non-rhetorical answer is that women are much more likely than men to join a Community Health team, for approximately the same reason that most companies’ HR teams are mostly women; and nobody has bothered to counteract this.
I had never considered that but I don’t think it’s a strong incentive. It doesn’t look like the Community Health team is doing this. If anything, I think they’re incentivized to give themselves less work, not more.
That’s not correct. Lots of EA orgs fundraise outside of the EA community.
Because guilt-by-association is a very weak form of argument. (And it’s not even obvious to me that there are relevant parallels there.) And FWIW I don’t respond to the sorts of people who use the word “TESCREAL” because I don’t think they’re worth taking seriously.
University groups do do those other things. But they do those things internally so you don’t notice. Recruiting is the only thing they do externally, so that’s what you notice.
Some orgs did that and it generally didn’t go well (eg Leverage Research). I think most people believe that totalizing jobs are bad for mental health and create bad epistemics and it’s not worth it.
Those are not examples of the unilateralist’s curse. I don’t want to explain it in this short comment but I would suggest re-reading some materials that explain the unilateralist’s curse, e.g. the original paper.
Because doing so would be a lot of work, which would take time away from doing other important things. I think people agree that having a second hub would be good, but not good enough to justify the effort.
Not necessarily guilt-by-association, but maybe rather pointing out that the two arguments/conspiracy theories share a similar flawed structure, so if you discredit one, you should discredit the other.
Still, I’m also unsure how much structure they share, and even if they did, I don’t think this would be discursively effective because I don’t think most people care that much about (that kind of) consistency (happy to be updated in the direction of most people caring about it).
Thanks for giving some answers here to these questions; it was really helpful to have them laid out like this.
1. In hindsight, I was probably talking more about moves towards decentralization of leadership, rather than decentralization of funding. I agree that greater decentralization of funding is a good thing, but it seems to me like, within the organizations funded by a given funder, decentralization of leadership is likely useless (if leadership decisions are still being made by informal networks between orgs rather than formal ones), or it may lead to a lack of clarity and direction.
3. I understand the dynamics that may cause the overrepresentation of women. However, that still doesn’t completely explain why there is an overrepresentation of white women, even when compared to racial demographics within EA at large. Additionally, this also doesn’t explain why the overrepresentation of women here isn’t seen as a problem on CEA’s part, if even just from an optics perspective.
4. Makes sense, but I’m still concerned that, say, if CEA had an anti-Stalinism team, they’d be reluctant to ever say “Stalinism isn’t a problem in EA.”
5. Again, this was a question that was badly worded on my end. I was referring more specifically to organizations within AI safety, more than EA at large. I know that AMF, GiveDirectly, The Humane League, etc. fundraise outside EA.
6. I was asking a descriptive question here, not a normative one. Guilt by association, even if weak, is a very commonly used form of argument, and so I would expect it to be in used in this case.
7. That makes sense. That was one of my hypotheses (hence my phrase “at least upon initial examination”), and I guess in hindsight it’s probably the best one.
10. Starting an AI capabilities company that does AI safety as a side project generally hasn’t gone well, and yet people keep doing it. The fact that something hasn’t gone well in the past doesn’t seem to me to be a sufficient explanation for why people don’t keep doing it, especially because it largely seems like Leverage failed for Leverage-specific reasons (i.e. too much engagement with woo). Additionally, your argument here seems to prove too much; the Manhattan Project was a large scientific project operating under an intense structure, and yet it was able to maintain good epistemics (i.e. not fixating too hard on designs that wouldn’t work) under those conditions. Same with a lot of really intense start-ups.
11. They may not be examples of the unilateralist’s curse in the original sense, but the term seems to have been expanded well past its original meaning, and they’re examples of that expanded meaning.
12. It seems to me like this is work of a different type than technical alignment work, and could likely be accomplished by hiring different people than the people already working on technical alignment, so it’s not directly trading off against that.
I intended my answer to be descriptive. EAs generally avoid making weak arguments (or at least I like to think we do).