X-risk, and AGI safety in particular, require unusual strength in gears-level reasoning to comprehend and work on; a willingness to stand up to criticism not only on technical questions but on moral/value questions; an intense, skeptical, questioning attitude; and a high value placed on altruism. Let’s call these people “rationalists.”
Even in scientific and engineering communities, and the population of rational people generally, the combination of these traits I’m referring to as “rationalism” is rare.
Rationalism causes people to have unusually high and predictable needs for a certain style and subject of debate and discourse, in a way that sets them apart from the general population.
Rationalists won’t be able to get their needs met in mainstream scientific or engineering communities, which prioritize a subset of the total rationalist package of traits.
Hence, they’ll seek an alternative community in which to get those needs met.
Rationalists who haven’t yet discovered a rationalist community won’t often have an advance knowledge of AGI safety. Instead, they’ll have thoughts and frustrations provoked by the non-rationalist society in which they grew up. It is these prosaic frustrations—often with politics—that will motivate them to seek out a different community, and to stay engaged with it.
When these people discover a community that engages with the controversial political topics they’ve seen shunned and censored in the rest of society, and doing it in a way that appears epistemically healthy to them, they’ll take it as evidence that they should stick around. It will also be a place where even AGI safety researchers and their friends can deal with the ongoing their issues and interests beyond AGI safety.
By associating with this community, they’ll pick up on ideas common in the community, like a concern for AGI safety. Some of them will turn it into a career, diminishing the amount of x-risk faced by the world.
I think that marginally increasing censorship on this site risks interfering with step 7. This site will not be recognized by proto-rationalists as a place where they can deal with the frustrations that they’re wrestling with when they first discover it. They won’t see an open attitude of free enquiry modeled, but instead see the same dynamics of fear-based censorship that they encounter almost everywhere else. Likewise, established AGI safety people and their friends will lose a space for free enquiry, a space for intellectual play and exploration that can be highly motivating. Loss of that motivation and appeal may interrupt the pipeline or staying power for people to work on X-risks of all kinds, including AGI safety.
Politics continues to affect people even after they’ve come to understand why it’s so frustrating, and having a minimal space to deal with it on this website seems useful to me. When you have very little of something, losing another piece of it feels like a pretty big deal.
When these people discover a community that engages with the controversial political topics they’ve seen shunned and censored in the rest of society, and doing it in a way that appears epistemically healthy to them, they’ll take it as evidence that they should stick around.
What has gone into forming this model? I only have one datapoint on this (which is myself). I stuck around because of the quality of discussion (people are making sense here!); I don’t think the content mattered. But I don’t have strong resistance to believing that this is how it works for other people.
I think if your model is applied to the politics ban, it would say that it’s also quite bad (maybe not as bad because most politics stuff isn’t as shunned and censored as social justice stuff)? If that’s true, how would you feel about restructuring rather than widening the censorship? Start allowing some political discussions (I also keep thinking about Wei Dai’s “it’ll go there eventually so we should practice” argument) but censor the most controversial social justice stuff. I feel like the current solution isn’t pareto optimal in the {epistemic health} x {safety against backlash} space.
Anecdotal, but about a year ago I committed to the rationalist community for exactly the reasons described. I feel more accepted in rationalist spaces than trans spaces, even though rationalists semi-frequently argue against the standard woke line and trans spaces try to be explicitly welcoming.
Just extrapolating from my own experience. For me, the content was important.
I think where my model really meets challenges is that clearly, the political content on LW has alienated some people. These people were clearly attracted here in the first place. My model says that LW is a magnet for likely AGI-safety researchers, and says nothing about it being a filter for likely AGI-safety researchers. Hence, if our political content is costing us more involvement than it’s retaining, or if the frustration experienced by those who’ve been troubled by the political content outweigh the frustration that would be experienced by those whose content would be censored, then that poses a real problem for my cost/benefit analysis.
A factor asymmetrically against increased censorship here is that censorship is, to me, intrinsically bad. It’s a little like war. Sometimes, you have to fight a war, but you should insist on really good evidence before you commit to it, because wars are terrible. Likewise, censorship sucks, and you should insist on really good evidence before you accept an increase in censorship.
It’s this factor, I think, that tilts me onto the side of preferring the present level of political censorship rather than an increase. I acknowledge and respect the people who feel they can’t participate here because they experience the environment as toxic. I think that is really unfortunate. I also think that censorship sucks, and for me, it roughly balances out with the suckiness of alienating potential participants via a lack of censorship.
This, I think, is the area where my mind is most susceptible to change. If somebody could make a strong case that LW currently has a lot of excessively toxic, alienating content, that this is the main bottleneck for wider participation, and that the number of people who’d leave if that controversial content were removed were outweighed by the number of people who’d join, then I’d be open-minded about that marginal increase in censorship.
An example of a way this evidence could be gathered would be some form of community outreach to ex-LWers and marginal LWers. We’d ask those people to give specific examples of the content they find offensive, and try both to understand why it bothers them, and why they don’t feel it’s something they can or want to tolerate. Then we’d try to form a consensus with them about limitations on political or potentially offensive speech that they would find comfortable, or at least tolerable. We’d also try to understand their level of interest in participating in a version of LW with more of these limitations in place.
Here, I am hypothesizing that there’s a group of ex-LWers or marginal LW-ers who feel a strong affinity for most of the content, while an even stronger aversion for a minority subset of the content to such a degree that they sharply curtail their participation. Such that if the offensive tiny fraction of the content were removed, they’d undergo a dramatic and lasting increase in engagement with LW. I find it unlikely that a sizeable group like this exists, but am very open to having my mind changed via some sort of survey data.
It seems more likely to me that ex/marginal-LWers are people with only a marginal interest in the site as a whole, who point to the minority of posts they find offensive as only the most salient example of what they dislike. Even if it were removed, they wouldn’t participate.
At the same time, we’d engage in community dialog with current active participants about their concerns with such a change. How strong are their feelings about such limitations? How many would likely stop reading/posting/commenting if these limitations were imposed? For the material they feel most strongly about it, why do they feel that way?
I am positing that there are a significant subset of LWers for whom the minority of posts engaging with politics are very important sources of its appeal.
How is it possible that I could simultaneously be guessing—and it is just a guess—that controversial political topics are a make-or-break screening-in feature, but not a make-or-break screening-out feature?
The reason is that there are abundant spaces online and in-person for conversation that does have the political limitations you are seeking to impose here. There are lots of spaces for conversation with a group of likeminded ideologues across the entire political spectrum, where conformity is a prerequisite of polite conversation. Hence, imposing the same sort of guardrails or ideological conformities on this website would make it similar to many other platforms. People who desire these guardrails/conformities can get what they want elsewhere. For them, LW would be a nice-to-have.
For those who desire polite and thoughtful conversation on a variety of intellectual topics, even touching on politics, LW is verging on a need-to-have. It’s rare. This is why I am guessing that a marginal increase in censorship would cost us more appeal than it would gain us.
I agree with you that the risk of being the subject of massive unwanted attention as a consequence is nonzero. I simply am guessing that it’s small enough not to be worth the ongoing short-term costs of a marginal increase in censorship.
But I do think that making the effort to thoroughly examine and gather evidence for the extent to which our political status quo serves to attract or repel people would be well worth a thorough examination. Asking at what point the inherent cost of a marginal increase in censorship becomes worth paying in exchange for a more inclusive environment seems like a reasonable question to ask. But I think this process would need a lot of community buy-in and serious effort on the part of a whole team to do it right.
The people who are already here would need persuading, and indeed, I think they deserve the effort to be persuaded to give up some of their freedom to post what they want here in exchange for, the hope would be, a larger and more vibrant community. And this effort should come with a full readiness to discover that, in fact, such restrictions would diminish the size and vibrancy and intellectual capacity of this community. If it wasn’t approached in that spirit, I think it would just fail.
Sorry for conflating anon03′s idea with yours!
A quick sketch at a gears-level model:
X-risk, and AGI safety in particular, require unusual strength in gears-level reasoning to comprehend and work on; a willingness to stand up to criticism not only on technical questions but on moral/value questions; an intense, skeptical, questioning attitude; and a high value placed on altruism. Let’s call these people “rationalists.”
Even in scientific and engineering communities, and the population of rational people generally, the combination of these traits I’m referring to as “rationalism” is rare.
Rationalism causes people to have unusually high and predictable needs for a certain style and subject of debate and discourse, in a way that sets them apart from the general population.
Rationalists won’t be able to get their needs met in mainstream scientific or engineering communities, which prioritize a subset of the total rationalist package of traits.
Hence, they’ll seek an alternative community in which to get those needs met.
Rationalists who haven’t yet discovered a rationalist community won’t often have an advance knowledge of AGI safety. Instead, they’ll have thoughts and frustrations provoked by the non-rationalist society in which they grew up. It is these prosaic frustrations—often with politics—that will motivate them to seek out a different community, and to stay engaged with it.
When these people discover a community that engages with the controversial political topics they’ve seen shunned and censored in the rest of society, and doing it in a way that appears epistemically healthy to them, they’ll take it as evidence that they should stick around. It will also be a place where even AGI safety researchers and their friends can deal with the ongoing their issues and interests beyond AGI safety.
By associating with this community, they’ll pick up on ideas common in the community, like a concern for AGI safety. Some of them will turn it into a career, diminishing the amount of x-risk faced by the world.
I think that marginally increasing censorship on this site risks interfering with step 7. This site will not be recognized by proto-rationalists as a place where they can deal with the frustrations that they’re wrestling with when they first discover it. They won’t see an open attitude of free enquiry modeled, but instead see the same dynamics of fear-based censorship that they encounter almost everywhere else. Likewise, established AGI safety people and their friends will lose a space for free enquiry, a space for intellectual play and exploration that can be highly motivating. Loss of that motivation and appeal may interrupt the pipeline or staying power for people to work on X-risks of all kinds, including AGI safety.
Politics continues to affect people even after they’ve come to understand why it’s so frustrating, and having a minimal space to deal with it on this website seems useful to me. When you have very little of something, losing another piece of it feels like a pretty big deal.
What has gone into forming this model? I only have one datapoint on this (which is myself). I stuck around because of the quality of discussion (people are making sense here!); I don’t think the content mattered. But I don’t have strong resistance to believing that this is how it works for other people.
I think if your model is applied to the politics ban, it would say that it’s also quite bad (maybe not as bad because most politics stuff isn’t as shunned and censored as social justice stuff)? If that’s true, how would you feel about restructuring rather than widening the censorship? Start allowing some political discussions (I also keep thinking about Wei Dai’s “it’ll go there eventually so we should practice” argument) but censor the most controversial social justice stuff. I feel like the current solution isn’t pareto optimal in the {epistemic health} x {safety against backlash} space.
Anecdotal, but about a year ago I committed to the rationalist community for exactly the reasons described. I feel more accepted in rationalist spaces than trans spaces, even though rationalists semi-frequently argue against the standard woke line and trans spaces try to be explicitly welcoming.
Just extrapolating from my own experience. For me, the content was important.
I think where my model really meets challenges is that clearly, the political content on LW has alienated some people. These people were clearly attracted here in the first place. My model says that LW is a magnet for likely AGI-safety researchers, and says nothing about it being a filter for likely AGI-safety researchers. Hence, if our political content is costing us more involvement than it’s retaining, or if the frustration experienced by those who’ve been troubled by the political content outweigh the frustration that would be experienced by those whose content would be censored, then that poses a real problem for my cost/benefit analysis.
A factor asymmetrically against increased censorship here is that censorship is, to me, intrinsically bad. It’s a little like war. Sometimes, you have to fight a war, but you should insist on really good evidence before you commit to it, because wars are terrible. Likewise, censorship sucks, and you should insist on really good evidence before you accept an increase in censorship.
It’s this factor, I think, that tilts me onto the side of preferring the present level of political censorship rather than an increase. I acknowledge and respect the people who feel they can’t participate here because they experience the environment as toxic. I think that is really unfortunate. I also think that censorship sucks, and for me, it roughly balances out with the suckiness of alienating potential participants via a lack of censorship.
This, I think, is the area where my mind is most susceptible to change. If somebody could make a strong case that LW currently has a lot of excessively toxic, alienating content, that this is the main bottleneck for wider participation, and that the number of people who’d leave if that controversial content were removed were outweighed by the number of people who’d join, then I’d be open-minded about that marginal increase in censorship.
An example of a way this evidence could be gathered would be some form of community outreach to ex-LWers and marginal LWers. We’d ask those people to give specific examples of the content they find offensive, and try both to understand why it bothers them, and why they don’t feel it’s something they can or want to tolerate. Then we’d try to form a consensus with them about limitations on political or potentially offensive speech that they would find comfortable, or at least tolerable. We’d also try to understand their level of interest in participating in a version of LW with more of these limitations in place.
Here, I am hypothesizing that there’s a group of ex-LWers or marginal LW-ers who feel a strong affinity for most of the content, while an even stronger aversion for a minority subset of the content to such a degree that they sharply curtail their participation. Such that if the offensive tiny fraction of the content were removed, they’d undergo a dramatic and lasting increase in engagement with LW. I find it unlikely that a sizeable group like this exists, but am very open to having my mind changed via some sort of survey data.
It seems more likely to me that ex/marginal-LWers are people with only a marginal interest in the site as a whole, who point to the minority of posts they find offensive as only the most salient example of what they dislike. Even if it were removed, they wouldn’t participate.
At the same time, we’d engage in community dialog with current active participants about their concerns with such a change. How strong are their feelings about such limitations? How many would likely stop reading/posting/commenting if these limitations were imposed? For the material they feel most strongly about it, why do they feel that way?
I am positing that there are a significant subset of LWers for whom the minority of posts engaging with politics are very important sources of its appeal.
How is it possible that I could simultaneously be guessing—and it is just a guess—that controversial political topics are a make-or-break screening-in feature, but not a make-or-break screening-out feature?
The reason is that there are abundant spaces online and in-person for conversation that does have the political limitations you are seeking to impose here. There are lots of spaces for conversation with a group of likeminded ideologues across the entire political spectrum, where conformity is a prerequisite of polite conversation. Hence, imposing the same sort of guardrails or ideological conformities on this website would make it similar to many other platforms. People who desire these guardrails/conformities can get what they want elsewhere. For them, LW would be a nice-to-have.
For those who desire polite and thoughtful conversation on a variety of intellectual topics, even touching on politics, LW is verging on a need-to-have. It’s rare. This is why I am guessing that a marginal increase in censorship would cost us more appeal than it would gain us.
I agree with you that the risk of being the subject of massive unwanted attention as a consequence is nonzero. I simply am guessing that it’s small enough not to be worth the ongoing short-term costs of a marginal increase in censorship.
But I do think that making the effort to thoroughly examine and gather evidence for the extent to which our political status quo serves to attract or repel people would be well worth a thorough examination. Asking at what point the inherent cost of a marginal increase in censorship becomes worth paying in exchange for a more inclusive environment seems like a reasonable question to ask. But I think this process would need a lot of community buy-in and serious effort on the part of a whole team to do it right.
The people who are already here would need persuading, and indeed, I think they deserve the effort to be persuaded to give up some of their freedom to post what they want here in exchange for, the hope would be, a larger and more vibrant community. And this effort should come with a full readiness to discover that, in fact, such restrictions would diminish the size and vibrancy and intellectual capacity of this community. If it wasn’t approached in that spirit, I think it would just fail.