Thanks for the detailed response! I’ll respond to a handful of points:
Previously “ignorant” people feel the community has opened a new world to them, they lived in darkness before, but now they found the “Way” (“Bayescraft”) and all this stuff is becoming an identity for them.
I certainly agree that there are people here who match that description, but it’s also worth pointing out that there are actual experts too.
the general public, who are just irrational automata still living in the dark.
One of the things I find most charming about LW, compared to places like RationalWiki, is how much emphasis there is on self-improvement and your mistakes, not mistakes made by other people because they’re dumb.
It seems that people try to prove they know some concept by using the jargon and including links to them. Instead, I’d prefer authors who actively try to minimize the need for links and jargon.
I’m not sure this is avoidable, and in full irony I’ll link to the wiki page that explains why.
In general, there are lots of concepts that seem useful, but the only way we have to refer to concepts is either to refer to a label or to explain the concept. A number of people read through the sequences and say “but the conclusions are just common sense!”, to which the response is, “yes, but how easy is it to communicate common sense?” It’s one thing to be able to recognize that there’s some vague problem, and another thing to be able to say “the problem here is inferential distance; knowledge takes many steps to explain, and attempts to explain it in fewer steps simply won’t work, and the justification for this potentially surprising claim is in Appendix A.” It is one thing to be able to recognize a concept as worthwhile; it is another thing to be able to recreate that concept when a need arises.
Now, I agree with you that having different labels to refer to the same concept, or conceptual boundaries or definitions that are drawn slightly differently, is a giant pain. When possible, I try to bring the wider community’s terminology to LW, but this requires being in both communities, which limits how much any individual person can do.
I also don’t get why the rationality stuff is intermixed with friendly AI and cryonics and transhumanism.
Part of that is just seeding effects—if you start a rationality site with a bunch of people interested in transhumanism, the site will remain disproportionately linked to transhumanism because people who aren’t transhumanists will be more likely to leave and people who are transhumanists will be more likely to find and join the site.
Part of it is that those are the cluster of ideas that seem weird but ‘hold up’ under investigation—most of the reasons to believe that the economy of fifty years from now will look like the economy of today are just confused, and if a community has good tools for dissolving confusions you should expect them to converge on the un-confused answer.
A final part seems to be availability; people who are convinced by the case for cryonics tend to be louder than the people who are unconvinced. The annual surveys show the perception of LW one gets from just reading posts (or posts and comments) is skewed from the perception of LW one gets from the survey results.
One of the things I find most charming about LW, compared to places like RationalWiki, is how much emphasis there is on self-improvement and your mistakes, not mistakes made by other people because they’re dumb.
I agree that LW is much better than RationalWiki, but I still think that the norms for discussion are much too far in the direction of focus on how other commenters are wrong as opposed to how one might oneself be wrong.
I know that there’s a selection effect (with respect to the more frustrating interactions standing out). But people not infrequently mistakenly believe that I’m wrong about things that I know much more about than they do, with very high confidence, and in such instances I find the connotations that I’m unsound to be exasperating.
I don’t think that this is just a problem for me rather than a problem for the community in general: I know a number of very high quality thinkers in real life who are uninterested in participating on LW explicitly because they don’t want to engage with commenters who are highly confident that their own positions are incorrect. There’s another selection effect here: such people aren’t salient because they’re invisible to the online community.
I know that there’s a selection effect (with respect to the more frustrating interactions standing out).
I agree that those frustrating interactions both happen and are frustrating, and that it leads to a general acidification of the discussion as people who don’t want to deal with it leave. Reversing that process in a sustainable way is probably the most valuable way to improve LW in the medium term.
Thanks for the detailed response! I’ll respond to a handful of points:
I certainly agree that there are people here who match that description, but it’s also worth pointing out that there are actual experts too.
One of the things I find most charming about LW, compared to places like RationalWiki, is how much emphasis there is on self-improvement and your mistakes, not mistakes made by other people because they’re dumb.
I’m not sure this is avoidable, and in full irony I’ll link to the wiki page that explains why.
In general, there are lots of concepts that seem useful, but the only way we have to refer to concepts is either to refer to a label or to explain the concept. A number of people read through the sequences and say “but the conclusions are just common sense!”, to which the response is, “yes, but how easy is it to communicate common sense?” It’s one thing to be able to recognize that there’s some vague problem, and another thing to be able to say “the problem here is inferential distance; knowledge takes many steps to explain, and attempts to explain it in fewer steps simply won’t work, and the justification for this potentially surprising claim is in Appendix A.” It is one thing to be able to recognize a concept as worthwhile; it is another thing to be able to recreate that concept when a need arises.
Now, I agree with you that having different labels to refer to the same concept, or conceptual boundaries or definitions that are drawn slightly differently, is a giant pain. When possible, I try to bring the wider community’s terminology to LW, but this requires being in both communities, which limits how much any individual person can do.
Part of that is just seeding effects—if you start a rationality site with a bunch of people interested in transhumanism, the site will remain disproportionately linked to transhumanism because people who aren’t transhumanists will be more likely to leave and people who are transhumanists will be more likely to find and join the site.
Part of it is that those are the cluster of ideas that seem weird but ‘hold up’ under investigation—most of the reasons to believe that the economy of fifty years from now will look like the economy of today are just confused, and if a community has good tools for dissolving confusions you should expect them to converge on the un-confused answer.
A final part seems to be availability; people who are convinced by the case for cryonics tend to be louder than the people who are unconvinced. The annual surveys show the perception of LW one gets from just reading posts (or posts and comments) is skewed from the perception of LW one gets from the survey results.
I agree that LW is much better than RationalWiki, but I still think that the norms for discussion are much too far in the direction of focus on how other commenters are wrong as opposed to how one might oneself be wrong.
I know that there’s a selection effect (with respect to the more frustrating interactions standing out). But people not infrequently mistakenly believe that I’m wrong about things that I know much more about than they do, with very high confidence, and in such instances I find the connotations that I’m unsound to be exasperating.
I don’t think that this is just a problem for me rather than a problem for the community in general: I know a number of very high quality thinkers in real life who are uninterested in participating on LW explicitly because they don’t want to engage with commenters who are highly confident that their own positions are incorrect. There’s another selection effect here: such people aren’t salient because they’re invisible to the online community.
I agree that those frustrating interactions both happen and are frustrating, and that it leads to a general acidification of the discussion as people who don’t want to deal with it leave. Reversing that process in a sustainable way is probably the most valuable way to improve LW in the medium term.