Assume Bad Faith
I’ve been trying to avoid the terms “good faith” and “bad faith”. I’m suspicious that most people who have picked up the phrase “bad faith” from hearing it used, don’t actually know what it means—and maybe, that the thing it does mean doesn’t carve reality at the joints.
People get very touchy about bad faith accusations: they think that you should assume good faith, but that if you’ve determined someone is in bad faith, you shouldn’t even be talking to them, that you need to exile them.
What does “bad faith” mean, though? It doesn’t mean “with ill intent.” Following Wikipedia, bad faith is “a sustained form of deception which consists of entertaining or pretending to entertain one set of feelings while acting as if influenced by another.” The great encyclopedia goes on to provide examples: the solider who waves a flag of surrender but then fires when the enemy comes out of their trenches, the attorney who prosecutes a case she knows to be false, the representative of a company facing a labor dispute who comes to the negotiating table with no intent of compromising.
That is, bad faith is when someone’s apparent reasons for doing something aren’t the same as the real reasons. This is distinct from malign intent. The uniformed solider who shoots you without pretending to surrender is acting in good faith, because what you see is what you get: the man whose clothes indicate that his job is to try to kill you is, in fact, trying to kill you.
The policy of assuming good faith (and mercilessly punishing rare cases of bad faith when detected) would make sense if you lived in an honest world where what you see generally is what you get (and you wanted to keep it that way), a world where the possibility of hidden motives in everyday life wasn’t a significant consideration.
On the contrary, however, I think hidden motives in everyday life are ubiquitous. As evolved creatures, we’re designed to believe as it benefited our ancestors to believe. As social animals in particular, the most beneficial belief isn’t always the true one, because tricking your conspecifics into adopting a map that implies that they should benefit you is sometimes more valuable than possessing the map that reflects the territory, and the most persuasive lie is the one you believe yourself. The universal human default is to come up with reasons to persuade the other party why it’s in their interests to do what you want—but admitting that you’re doing that isn’t part of the game. A world where people were straightforwardly trying to inform each other would look shocking and alien to us.
But if that’s the case (and you shouldn’t take my word for it), being touchy about bad faith accusations seems counterproductive. If it’s common for people’s stated reasons to not be the same as the real reasons, it shouldn’t be beyond the pale to think that of some particular person, nor should it necessarily entail cutting the “bad faith actor” out of public life—if only because, applied consistently, there would be no one left. Why would you trust anyone so highly as to think they never have a hidden agenda? Why would you trust yourself?
The conviction that “bad faith” is unusual contributes to a warped view of the world in which conditions of information warfare are rationalized as an inevitable background fact of existence. In particular, people seem to believe that persistent good faith disagreements are an ordinary phenomenon—that there’s nothing strange or unusual about a supposed state of affairs in which I’m an honest seeker of truth, and you’re an honest seeker of truth, and yet we end up persistently disagreeing on some question of fact.
I claim that this supposedly ordinary state of affairs is deeply weird at best, and probably just fake. Actual “good faith” disagreements—those where both parties are just trying to get the right answer and there are no other hidden motives, no “something else” going on—tend not to persist.
If this claim seems counterintuitive, you may not be considering all the everyday differences in belief that are resolved so quickly and seamlessly that we tend not to notice them as “disagreements”.
Suppose you and I have been planning to go to a concert, which I think I remember being on Thursday. I ask you, “Hey, the concert is on Thursday, right?” You say, “No, I just checked the website; it’s on Friday.”
In this case, I immediately replace my belief with yours. We both just want the right answer to the factual question of when the concert is. With no “something else” going on, there’s nothing stopping us from converging in one step: your just having checked the website is a more reliable source than my memory, and neither you nor the website have any reason to lie. Thus, I believe you; end of story.
In cases where the true answer is uncertain, we expect similarly quick convergence in probabilistic beliefs. Suppose you and I are working on some physics problem. Both of us just want the right answer, and neither of us is particularly more skilled than the other. As soon as I learn that you got a different answer than me, my confidence in my own answer immediately plummets: if we’re both equally good at math, then each of us is about as likely to have made a mistake. Until we compare calculations and work out which one of us (or both) made a mistake, I think you’re about as likely to be right as me, even if I don’t know how you got your answer. It wouldn’t make sense for me to bet money on my answer being right simply because it’s mine.
Most disagreements of note—most disagreements people care about—don’t behave like the concert date or physics problem examples: people are very attached to “their own” answers. Sometimes, with extended argument, it’s possible to get someone to change their mind or admit that the other party might be right, but with nowhere near the ease of agreeing on (probabilities of) the date of an event or the result of a calculation—from which we can infer that, in most disagreements people care about, there is “something else” going on besides both parties just wanting to get the right answer.
But if there’s “something else” going on in typical disagreements that look like a grudge match rather than a quick exchange of information resulting in convergence of probabilities, then the belief that persistent good faith disagreements are common would seem to be in bad faith! (Because if bad faith is “entertaining [...] one set of feelings while acting as if influenced by another”, believers in persistent good faith disagreements are entertaining the feeling that both parties to such a disagreement are honest seekers of truth, but acting otherwise insofar as they anticipate seeing a grudge match rather than convergence.)
Some might object that bad faith is about conscious intent to deceive: honest reporting of unconsciously biased beliefs isn’t bad faith. I’ve previously expressed doubt as to how much of what we call lying requires conscious deliberation, but a more fundamental reply is that from the standpoint of modeling information transmission, the difference between bias and deception is uninteresting—usually not relevant to what probability updates should be made.
If an apple is green, and you tell me that it’s red, and I believe you, I end up with false beliefs about the apple. It doesn’t matter whether you said it was red because you were consciously lying or because you’re wearing rose-colored glasses. The input–output function is the same either way: the problem is that the color you report to me doesn’t depend on the color of the apple.
If I’m just trying to figure out the relationship between your reports and the state of the world (as contrasted to caring about punishing liars while letting merely biased people off the hook), the main reason to care about the difference between unconscious bias and conscious deception is that the latter puts up much stronger resistance. Someone who is merely biased will often fold when presented with a sufficiently compelling counterargument (or reminded to take off their rose-colored glasses); someone who’s consciously lying will keep lying (and telling ancillary lies to cover up the coverup) until you catch them red-handed in front of an audience with power over them.
Given that there’s usually “something else” going on in persistent disagreements, how do we go on, if we can’t rely on the assumption of good faith? I see two main strategies, each with their own cost–benefit profile.
One strategy is to stick the object level. Arguments can be evaluated on their merits, without addressing what the speaker’s angle is in saying it (even if you think there’s probably an angle). This delivers most of the benefits of “assume good faith” norms; the main difference I’m proposing is that speakers’ intentions be regarded as off-topic rather than presumed to be honest.
The other strategy is full-contact psychoanalysis: in addition to debating the object-level arguments, interlocutors have free reign to question each other’s motives. This is difficult to pull off, which is why most people most of the time should stick to the object level. Done well, it looks like a negotiation: in the course of discussion, pseudo-disagreements (where I argue for a belief because it’s in my interests for that belief to be on the shared map) are factorized out into real disagreements and bargaining over interests so that Pareto improvements can be located and taken, rather than both parties fighting to distort the shared map in the service of their interests.
For an example of what a pseudo-disagreement looks like, imagine that I own a factory that I’m considering expanding onto the neighboring wetlands, and you run a local environmental protection group. The regulatory commission with the power to block the factory expansion has a mandate to protect local avian life, but not to preserve wetland area. The factory emits small amounts of Examplene gas. You argue before the regulatory commission that the expansion should be blocked because the latest Science shows that Examplene makes birds sad. I counterargue that the latest–latest Science shows that Examplene actually makes birds happy; the previous studies misheard their laughter as tears and should be retracted.
Realistically, it seems unlikely that our apparent disagreement is “really” about the effects of Examplene on avian mood regulation. More likely, what’s actually going on is a conflict rather than a disagreement: I want to expand my factory onto the wetlands, and you want me to not do that. The question of how Examplene pollution affects birds only came into it in order to persuade the regulatory commission.
It’s inefficient that our conflict is being disguised as a disagreement. We can’t both get what we want, but however the factory expansion question ultimately gets resolved, it would be better to reach that outcome without distorting Society’s shared map of the bioactive properties of Examplene. (Maybe it doesn’t affect the birds at all!) Whatever the true answer is, Society has a better shot at figuring it out if someone is allowed to point out your bias and mine (because facts about which evidence gets promoted to one’s attention are relevant to how one should update on that evidence).
The reason I don’t think it’s useful to talk about “bad faith” is because the ontology of good vs. bad faith isn’t a great fit to either discourse strategy.
If I’m sticking to the object level, it’s irrelevant: I reply to what’s in the text; my suspicions about the process generating the text are out of scope.
If I’m doing full-contact psychoanalysis, the problem with “I don’t think you’re here in good faith” is that it’s insufficiently specific. Rather than accusing someone of generic “bad faith”, the way to move the discussion forward is by positing that one’s interlocutor has some specific motive that hasn’t yet been made explicit—and the way to defend oneself against such an accusation is by making the case that one’s real agenda isn’t the one being proposed, rather than protesting one’s “good faith” and implausibly claiming not to have an agenda.
The two strategies can be mixed. A simple meta-strategy that performs well without imposing too high of a skill requirement is to default to the object level, and only pull out psychoanalysis as a last resort against stonewalling.
Suppose you point out that my latest reply seems to contradict something I said earlier, and I say, “Look over there, a distraction!”
If you want to continue sticking to the object level, you could say, “I don’t understand how the distraction is relevant to resolving the inconsistency in your statements that I raised.” On the other hand, if you want to drop down into psychoanalysis, you could say, “I think you’re only pointing out the distraction because you don’t want to be pinned down.” Then I would be forced to either address your complaint, or explain why I had some other reason to point out the distraction.
Crucially, however, the choice of whether to investigate motives doesn’t depend on an assumption that only “bad guys” have motives—as if there were bad faith actors who have an angle, and good faith actors who are ideal philosophers of perfect emptiness. There’s always an angle; the question is which one.
- And All the Shoggoths Merely Players by 10 Feb 2024 19:56 UTC; 170 points) (
- Voting Results for the 2023 Review by 6 Feb 2025 8:00 UTC; 86 points) (
- 31 Aug 2023 18:39 UTC; 53 points) 's comment on Introducing the Center for AI Policy (& we’re hiring!) by (
- On the Contrary, Steelmanning Is Normal; ITT-Passing Is Niche by 9 Jan 2024 23:12 UTC; 45 points) (
- 22 Sep 2023 20:07 UTC; 6 points) 's comment on Would You Work Harder In The Least Convenient Possible World? by (
- 22 Feb 2024 20:25 UTC; 1 point) 's comment on mike_hawke’s Shortform by (
- 15 Sep 2023 21:51 UTC; 1 point) 's comment on Contra Yudkowsky on Epistemic Conduct for Author Criticism by (
(Self-review.) I claim that this post is significant for articulating a solution to the mystery of disagreement (why people seem to believe different things, in flagrant violation of Aumann’s agreement theorem): much of the mystery dissolves if a lot of apparent “disagreements” are actually disguised conflicts. The basic idea isn’t particularly original, but I’m proud of the synthesis and writeup. Arguing that the distinction between deception and bias is less decision-relevant than commonly believed seems like an improvement over hang-wringing over where the boundary is.
I think that a lot of disagreements are truly hidden conflicts, but I also do think that a non-trivial portion of disagreements come down to not having common priors, which is necessary for the theorem to work for disagreements.
I think this post is quite good, and gives a heuristic important to modeling the world. If you skipped it because of title + author, you probably have the wrong impression of its contents and should give it a skim. Its main problem is what’s left unsaid.
Some people in the comments reply to it that other people self-deceive, yes, but you should assume good faith. I say—why not assume the truth, and then do what’s prosocial anyways?
One reason to assume good faith is to stop yourself from justifying your own hidden motives in a conflict:
https://www.lesswrong.com/posts/e4GBj6jxRZcsHFSvP/assume-bad-faith#Sc6RqbDpurX6hJ8pY
+4. I most like the dichotomoy of “stick to object level” vs “full contact psychoanalysis.” And I think the paragraphs towards the end are important:
I think maybe the title of this post is misleading, and sort of clickbaity/embroiled in a particular conflict that is, ironically, a distraction.
The whole post (admittedly from the very first sentence) is actually about avoiding the frame of bad faith (even if the post argues that basically everyone is in bad faith most of the time). I think it’s useful if post titles convey more of a pointer to the core idea of the post, which gives it a shorthand that’s more likely to be remembered/used when the time is right. (“Taboo bad faith”? “The object level, vs full-contact psychanalysis”? Dunno. Both of those feel like they lose some nuance I think Zack cares about. But, I think there’s some kind of improvement here)
...
That said: I think the post is missing something that lc’s comment sort of hints at (although I don’t think lc meant to imply my takeaway)
I think the post does give a couple concrete strategies for how to navigate the, er, conflict between conflict and truthseeking, which are one flavor of “prosocial.” I think it’s missing something about what “assume good faith” is for, which isn’t really covered in the post.
The problem “assume good faith” is trying to solve is “there are a bunch of human tendencies to get into escalation spirals of distrust, and make a bigger deal about the mistakes of people you’re in conflict with.” You don’t have to fix this with false beliefs, you can fix this by shifting your relational stance towards the person and the discussion, and holding the possibility more alive that you’ve misinterpreted them or are rounding them off to and incorrect guess as to what their agenda is.
I think the suggestions this post makes on how to deal with that are reasonable, but incomplete, and I think people benefit from also having some tools that more directly engage with “your impulse to naively assume bad faith is part of a spirally-pattern that you may want to step out of somehow.”
The “Taboo bad faith” title doesn’t fit this post. I had hoped from the opening section that it was going in that direction, but it did not.
Most obviously, the post kept relying heavily on the terms “bad faith” and “good faith” and that conceptual distinction, rather than tabooing them.
But also, it doesn’t do the core intellectual work of replacing a pointer with its substance. In the opening scenario where someone accuses their conversation partner of bad faith, conveying something along the lines of ‘I disapprove of how you’re approaching this conversation so I’m leaving’, tabooing “bad faith” would mean articulating what pattern of behavior (they thought that) they saw and why disapproval & departure is an appropriate response. Zack doesn’t try to do this, he just abandons this scenario to talk about other things involving his definition of “bad faith”. (And similarly with “assume good faith”.) I briefly hoped that the post would go in the “taboo your words” direction, describing what was happening in that sort of scenario with a clarity and precision that would make the label “bad faith” seem crude by comparison, but it did not.
This post also doesn’t manage to avoid the main pitfall that tabooing a word is meant to prevent, where people talk past each other because they’re using the same word with different definitions. Even though he says at the start of the post that other people are using the term “bad/good faith” wrong according his understanding of the term, when he talks about the advice “assume good faith” he just plugs in his definition of “good faith” (and “assume”) without noting that he’s making an interpretation of what other people mean when they use the phrase and that they might mean something else. And similarly in other places like “being touchy about bad faith accusations seems counterproductive” and “the belief that persistent good faith disagreements are common would seem to be in bad faith”. When someone says “you’re acting in bad faith” are they claiming that you’re showing the thing that Zack means by “bad faith”? Keeping that sort of thing straight is rationality 101 stuff that tabooing words helps with, and which this post repeatedly stumbles over.
I am sort of confused at what you got from this post. You say “But also, it doesn’t do the core intellectual work of replacing a pointer with its substance.”, but, I think he explicitly does that here?
It feels like you wanted some entirely different post, about how to navigate when someone accuses someone of bad faith, for a variety of possible definitions of bad faith? (As opposed to this post, which is mostly saying “Avoid accusing people of bad faith. Instead, do some kind of more specific and useful thing.” Which honestly seems like good advice to me even for most people who are using the phrase to mean something else. “I disapprove of what you’re doing here and am leaving now” seems totally fine)
This post begins:
The second paragraph uses the term “bad faith” or “good faith” three times. What substance is it pointing to?
AFAICT the post never fleshes this out. The ‘hidden motives’ definition that Zack gave fleshes out his understanding of the term, which is different from what these people mean.
Tabooing words, when different people are using the word differently, typically means giving substance to both meanings (e.g. “acoustic vibrations” and “auditory experiences” for sound).
If Zack wanted to set aside the question of what other people mean by “bad faith” and just think about some things using his understanding of the term, then he could’ve done that. (To me that seems less interesting than also engaging with what other people mean by the term, and it would’ve made it a bit strange to start the post this way, but it still seems like a fine direction to go.) That’s not what this post did, though. It keeps coming back to what other people think about bad faith, without tracking that there are different meanings.
Consider this from Zack: “The conviction that “bad faith” is unusual contributes to a warped view of the world”. This is more on the topic of what other people think about “bad faith”. Which meaning of “bad faith” is it using? If it means Zack’s ‘hidden motives’ definition then it’s unclear if people do have the conviction that that’s unusual, because when people use the words “bad faith” that’s not what they’re talking about. If it means whatever people do mean by the words “bad faith”, then we’re back to discussing some substance that hasn’t been fleshed out, and it’s unclear if their conviction that it’s rare contributes to a warped view of the world because it’s unclear what that conviction even is.
Okay I think I have more of an idea where you’re coming from. (although I get some sense of something being at stake for you here that I still don’t understand).
I maybe want to clarify, when I suggested “taboo bad faith” as a title, I was imagining a pretty different purpose for the title (and I don’t super strongly defend that title as a good one here). I was looking for a succinct way of describing the suggestion “when you want to accuse someone of bad faith, you should probably say something more specific instead.” (i.e. “Taboo Bad Faith” is a recommendation for what to do in the wild, rather than “a thing the post itself was doing.”)