ITT-passing and civility are good; âcharityâ is bad; steelmanning is niche
This post has been recorded as part of the LessWrong Curated Podcast, and can be listened to on Spotify, Apple Podcasts, and Libsyn.
I often object to claims like âcharity/âsteelmanning is an argumentative virtueâ. This post collects a few things I and others have said on this topic over the last few years.
My current view is:
Steelmanning (âthe art of addressing the best form of the other personâs argument, even if itâs not the one they presentedâ) is a useful niche skill, but I donât think it should be a standard thing you bring out in most arguments, even if itâs an argument with someone you strongly disagree with.
Instead, arguments should mostly be organized around things like:
Object-level learning and truth-seeking, with the conversation as a convenient excuse to improve your own model of something youâre curious about.
Trying to pass each otherâs Ideological Turing Test (ITT), or some generalization thereof. The ability to pass ITTs is the ability âto state opposing views as clearly and persuasively as their proponentsâ.
The version of âITTâ I care about is one where you understand the substance of someoneâs view well enough to be able to correctly describe their beliefs and reasoning; I donât care about whether you can imitate their speech patterns, jargon, etc.
Trying to identify and resolve cruxes: things that would make one or the other of you (or both) change your mind about the topic under discussion.
Argumentative charity is a complete mess of a conceptâ âpeople use it to mean a wide variety of things, and many of those things are actively bad, or liable to cause severe epistemic distortion and miscommunication.
Some version of civility and/âor friendliness and/âor a spirit of camaraderie and goodwill seems like a useful ingredient in many discussions. Iâm not sure how best to achieve this in ways that are emotionally honest (âpretending to be cheerful and warm when you donât feel that wayâ sounds like the wrong move to me), or how to achieve this without steering away from candor, openness, ârealnessâ, etc.
Iâve said that I think people should be ânicer and also ruderâ. And:
The sweet spot for EA PR is something like: âfriendly, nuanced, patient, and totally unapologetic about being a fire hose of inflammatory hot takesâ. đ
I have an intuition that those are pieces of the puzzle, along with (certain aspects or interpretations of) NVC tech, circling tech, introspection tech, etc. But Iâm not sure how to hit the right balance in general.
I do feel very confident that âsteelmanningâ and âcharityâ arenât the right tech for achieving this goal. (Because âcharityâ is a bad meme, and âsteelmanningâ is a lot more niche than that.)
Things other people have said
Ozy Brennan wrote Against Steelmanning in 2016, and Eliezer Yudkowsky commented:
Be it clear: Steelmanning is not a tool of understanding and communication. The communication tool is the Ideological Turing Test. âSteelmanningâ is what you do to avoid the equivalent of dismissing AGI after reading a media argument. It usually indicates that you think youâre talking to somebody as hapless as the media.
The exception to this rule is when you communicate, âWell, on my assumptions, the plausible thing that sounds most like this is...â which is a cooperative way of communicating to the person what your own assumptions are and what you think are the strong and weak points of what you think might be the argument.
Mostly, you should be trying to pass the Ideological Turing Test if speaking to someone you respect, and offering âMy steelman might be...?â only to communicate your own premises and assumptions. Or maybe, if you actually believe the steelman, say, âI disagree with your reason for thinking X, but Iâll grant you X because I believe this other argument Y. Is that good enough to move on?â Be ready to accept âNo, the exact argument for X is important to my later conclusionsâ as an answer.
âLet me try to imagine a smarter version of this stupid positionâ is when youâve been exposed to the Deepak Chopra version of quantum mechanics, and you donât know if itâs the real version, or what a smart person might really think is the issue. Itâs what you do when you donât want to be that easily manipulated sucker who can be pushed into believing X by a flawed argument for not-X that you can congratulate yourself for being skeptically smarter than. Itâs not what you do in a respectful conversation.
In 2017, Holden Karnofsky wrote:
I try to avoid straw-manning, steel-manning, and nitpicking. I strive for an accurate understanding of the most important premises behind someoneâs most important decisions, and address those. (As a side note, I find it very unsatisfying to engage with âsteel-manâ versions of my arguments, which rarely resemble my actual views.)
And Eliezer wrote, in a private Facebook thread:
Reminder: Eliezer and Holden are both on record as saying that âsteelmanningâ people is bad and you should stop doing it.
As Holden says, if youâre trying to understand someone or you have any credence at all that they have a good argument, focus on passing their Ideological Turing Test. âSteelmanningâ usually ends up as weakmanning by comparison. If they donât in fact have a good argument, itâs falsehood to pretend they do. If you want to try to make a genuine effort to think up better arguments yourself because they might exist, donât drag the other person into it.
Things Iâve said
In 2018, I wrote:
When someone makes a mistake or has a wrong belief, you shouldnât âsteelmanâ that belief by replacing it with a different one; it makes it harder to notice mistakes and update from them, and it also makes it harder to understand peopleâs real beliefs and actions.
âWhat belief does this person have?â is a particular factual question. Steelmanning, like âcharityâ, is sort of about unfocusing your eyes and tricking yourself into treating the factual question as though it were a game: you want to find a fairness-preserving allocation of points to all players, where more credible views warrant more points. Some people like that act of unfocusing because itâs fun to brainstorm new arguments; or they think itâs a useful trick for reducing social conflict or resistance to new ideas. But itâs dangerous to frame that unfocusing as âsteelmanningâ or âcharityâ rather than explicitly flagging âI want to change the topic to this other thing your statement happened to remind me ofâ.
In 2019, I said:
Charity seems more useful for rhetoric/âpersuasion/âdiplomacy; steel-manning seems more useful for brainstorming; both seem dangerous insofar as they obscure the original meaning and make it harder to pass someoneâs Ideological Turing Test.
âCharityâ seems like the more dangerous meme to me because it encourages more fuzziness about whether youâre flesh-manning [i.e., just trying to accurately model] vs. steel-manning the argument, and because it has more moral overtones. Itâs more epistemically dangerous to filter your answers to factual questions by criteria other than truth, than to decide to propose a change of topic.
[...] I endorse ânon-uncharitablenessââtrying to combat biases toward having an inaccurately negative view of your political enemies and so on.
I worry that removing the double negative makes it seem like charity is an epistemic end in its own right, rather than an attempt to combat a bias. I also worry that the word âcharityâ makes it tempting to tie non-uncharitableness to niceness/âfriendliness, which makes it more effortful to think about and optimize those goals separately.
Most of my worries about charity and steelmanning go away if theyâre discussed with the framings ânon-uncharitableness and niceness are two separate goalsâ and âgood steelmanning and good fleshmanning are two separate goalsâ, respectively.
E.g., actively focus on examples of:
being epistemically charitable in ways that arenât nice, friendly, or diplomatic.
being nice and prosocial in ways that require interpreting the person as saying something less plausible.
trying to better pass someoneâs Ideological Turing Test by focusing on less plausible claims and arguments.
coming up with steelmen that explicitly assert the falsehood of the claim theyâre the steelman of.
I also think that the equivocation in âcharityâ is doing some conversational work.
E.g.: Depending on context and phrasing, saying that youâre optimizing for friendliness can make you seem manipulative or inauthentic, or it can seem like a boast or a backhanded attack (âI was trying to be nice when I said it that wayâ /â âIâm trying to be friendlyâ.) Framing a diplomatic goal as though it were epistemic can mitigate that problem.
Similarly, if youâre in an intellectual or academic environment and you want to criticize someone for being a jerk, âyouâre being uncharitableâ is likely to get less pushback, not only because itâs relatively dry but because criticisms of tone are generally more controversial among intellectuals than criticisms of content.
âYouâre being uncharitableâ is also a common accusation in a motte-and-bailey context. Any argument can be quickly dismissed if it makes your conclusion sound absurd, because the arguer must just be violating the principle of charity. It may not even be necessary to think of an alternative, stronger version of the claim under attack, if youâre having an argument over twitter and can safely toss out the âThat sounds awfully uncharitableâ line and then disappear in the mist.
⌠Hm, this comment ended up going in a more negative direction than I was intending. The concerns above are important, but the thing I originally intended to say was that itâs not an accident âcharityâ is equivocal, and thereâs some risk in disambiguating it without recognizing the conversational purposes the ambiguity was serving, contra my earlier insistence on burning the whole thing down. It may be helping make a lot of social interactions smoother, helping giving people more cover to drop false views with minimal embarrassment (by saying they really meant the more-charitable interpretation all along), etc.
(I now feel more confident that, no, âcharityâ is just a bad meme. Ditch it and replace it with something new.)
From 2021:
The problem isnât âcharity is a good conversational norm, but these people are doing it wrongâ; the problem is that charity is a bad conversational norm. If nothing else, itâs bad because it equivocates between âbe friendlyâ norms and âhave accurate beliefs about othersâ norms.
Good norms:
Keep discussions civil and chill.
Be wary of biases to strawman others.
Try to pass othersâ ITT.
Use steelmen to help you think outside the box.
Bad norms:
Treat the above norms as identical.
Try to delude yourself about how good othersâ arguments are.
From 2022:
I think the term âcharityâ is genuinely ambiguous about whether youâre trying to find the personâs true view, vs. trying to steel-man, vs. some combination. Different people at different times do all of those things and call it argumentative âcharityâ.
This if anything strikes me as even worse than saying âIâm steel-manningâ, because at least steel-manning is transparent about what itâs doing, even if people tend to underestimate the hazards of doing it.
- CritÂiÂcism of EA CritÂiÂcism Contest by 14 Jul 2022 14:30 UTC; 108 points) (
- CritÂiÂcism of EA CritÂiÂcism Contest by 14 Jul 2022 14:20 UTC; 101 points) (EA Forum;
- Notes on how I want to hanÂdle criticism by 8 Jun 2023 11:47 UTC; 63 points) (EA Forum;
- VotÂing ReÂsults for the 2022 Review by 2 Feb 2024 20:34 UTC; 57 points) (
- 22 Nov 2022 22:08 UTC; 55 points) 's comment on EA should blurt by (EA Forum;
- On the ConÂtrary, SteelÂmanÂning Is NorÂmal; ITT-PassÂing Is Niche by 9 Jan 2024 23:12 UTC; 44 points) (
- In faÂvor of steelmanning by 1 May 2023 17:12 UTC; 36 points) (
- In faÂvor of steelmanning by 1 May 2023 15:33 UTC; 27 points) (EA Forum;
- 20 Jan 2023 21:51 UTC; 6 points) 's comment on FLI FAQ on the reÂjected grant proÂposal controversy by (EA Forum;
- 5 Jun 2023 8:43 UTC; 4 points) 's comment on Tim BernÂers-Lee found it hard to exÂplain the web by (
Reply: âOn the Contrary, Steelmanning Is Normal; ITT-Passing Is Nicheâ