Pluralistic Moral Reductionism
Part of the sequence: No-Nonsense Metaethics
Disputes over the definition of morality… are disputes over words which raise no really significant issues. [Of course,] lack of clarity about the meaning of words is an important source of error… My complaint is that what should be regarded as something to be got out of the way in the introduction to a work of moral philosophy has become the subject matter of almost the whole of moral philosophy...
Peter Singer
If a tree falls in the forest, and no one hears it, does it make a sound? If by ‘sound’ you mean ‘acoustic vibrations in the air’, the answer is ‘Yes.’ But if by ‘sound’ you mean an auditory experience in the brain, the answer is ‘No.’
We might call this straightforward solution pluralistic sound reductionism. If people use the word ‘sound’ to mean different things, and people have different intuitions about the meaning of the word ‘sound’, then we needn’t endlessly debate which definition is ‘correct’.1 We can be pluralists about the meanings of ‘sound’.
To facilitate communication, we can taboo and reduce: we can replace the symbol with the substance and talk about facts and anticipations, not definitions. We can avoid using the word ‘sound’ and instead talk about ‘acoustic vibrations’ or ‘auditory brain experiences.’
Still, some definitions can be wrong:
Alex: If a tree falls in the forest, and no one hears it, does it make a sound?
Austere MetaAcousticist: Tell me what you mean by ‘sound’, and I will tell you the answer.
Alex: By ‘sound’ I mean ‘acoustic messenger fairies flying through the ether’.
Austere MetaAcousticist: There’s no such thing. Now, if you had asked me about this other definition of ‘sound’...
There are other ways for words to be wrong, too. But once we admit to multiple potentially useful reductions of ‘sound’, it is not hard to see how we could admit to multiple useful reductions of moral terms.
Many Moral Reductionisms
Moral terms are used in a greater variety of ways than sound terms are. There is little hope of arriving at the One True Theory of Morality by analyzing common usage or by triangulating from the platitudes of folk moral discourse. But we can use stipulation, and we can taboo and reduce. We can use pluralistic moral reductionism2 (for austere metaethics, not for empathic metaethics).
Example #1:
Neuroscientist Sam Harris: Which is better? Religious totalitarianism or the Northern European welfare state?
Austere Metaethicist: What do you mean by ‘better’?
Harris: By ‘better’ I mean ‘that which tends to maximize the well-being of conscious creatures’.
Austere Metaethicist: Assuming we have similar reductions of ‘well-being’ and ‘conscious creatures’ in mind, the evidence I know of suggests that the Northern European welfare state is more likely to maximize the well-being of conscious creatures than religious totalitarianism.
Example #2:
Philosopher Peter Railton: Is capitalism the best economic system?
Austere Metaethicist: What do you mean by ‘best’?
Railton: By ‘best’ I mean ‘would be approved of by an ideally instrumentally rational and fully informed agent considering the question ‘How best to maximize the amount of non-moral goodness?’ from a social point of view in which the interests of all potentially affected individuals are counted equally.
Austere Metaethicist: Assuming we agree on the meaning of ‘ideally instrumentally rational’ and ‘fully informed’ and ‘agent’ and ‘non-moral goodness’ and a few other things, the evidence I know of suggests that capitalism would not be approved of by an ideally instrumentally rational and fully informed agent considering the question ‘How best to maximize the amount of non-moral goodness?’ from a social point of view in which the interests of all potentially affected individuals were counted equally.
Example #3:
Theologian Bill Craig: Ought we to give 50% of our income to efficient charities?
Austere Metaethicist: What do you mean by ‘ought’?
Craig: By ‘ought’ I mean ‘approved of by an essentially just and loving God’.
Austere Metaethicist: Your definition doesn’t connect to reality. It’s like talking about atom-for-atom ‘indexical identity’ even though the world is made of configurations and amplitudes instead of Newtonian billiard balls. Gods don’t exist.
But before we get to empathic metaethics, let’s examine the standard problems of metaethics using the framework of pluralistic moral reductionism.
Cognitivism vs. Noncognitivism
One standard debate in metaethics is cognitivism vs. noncognitivism. Alexander Miller explains:
Consider a particular moral judgement, such as the judgement that murder is wrong. What sort of psychological state does this express? Some philosophers, called cognitivists, think that a moral judgement such as this expresses a belief.
Beliefs can be true or false: they are truth-apt, or apt to be assessed in terms of truth and falsity. So cognitivists think that moral judgements are capable of being true or false.
On the other hand, non-cognitivists think that moral judgements express non-cognitive states such as emotions or desires. Desires and emotions are not truth-apt. So moral judgements are not capable of being true or false.3
But why should we expect all people to use moral judgments like “Stealing is wrong” to express the same thing?4
Some people who say “Stealing is wrong” are really just trying to express emotions: “Stealing? Yuck!” Others use moral judgments like “Stealing is wrong” to express commands: “Don’t steal!” Still others use moral judgments like “Stealing is wrong” to assert factual claims, such as “stealing is against the will of God” or “stealing is a practice that usually adds pain rather than pleasure to the world.”
It may be interesting to study all such uses of moral discourse, but this post focuses on addressing cognitivists, who use moral judgments to assert factual claims. We ask: Are those claims true or false? What are their implications?
Objective vs. Subjective Morality
Is morality objective or subjective? It depends which moral reductionism you have in mind, and what you mean by ‘objective’ and ‘subjective’.
Here are some common5 uses of the objective/subjective distinction in ethics:
Moral facts are objective1 if they are made true or false by mind-independent facts, otherwise they are subjective1.
Moral facts are objective2 if they are made true or false by facts independent of the opinions of sentient beings, otherwise they are subjective2.
Moral facts are objective3 if they are made true or false by facts independent of the opinions of humans, otherwise they are subjective3.
Now, consider Harris’ reduction of morality to facts about the well-being of conscious creatures. His theory of morality is objective3 and objective2, because facts about well-being are independent of anyone’s opinion. Even if the Nazis had won WWII and brainwashed everybody to have the opinion that torturing Jews was moral, it would remain true that torturing Jews does not increase the average well-being of conscious creatures. But Harris’ theory of morality is not objective1, because facts about the well-being of conscious creatures are mind-dependent facts.
Or, consider Craig’s theory of morality in terms of divine approval. His theory doesn’t connect to reality, but still: is it objective or subjective? Craig’s theory says that moral facts are objective3, because they don’t depend on human opinion (God isn’t human). But his theory doesn’t say that morality is objective2 or objective1, because for him, moral facts depend on the opinion of a sentient being: God.
A warning: ambiguous terms like ‘objective’ and ‘subjective’ are attractors for sneaking in connotations. Craig himself provides an example. In his writings and public appearances, Craig insists that only God-based morality can be objective.6 What does he mean by ‘objective’? On a single page,7 he uses ‘objective’ to mean “independent of people’s opinions” (objective2) and also to mean “independent of human opinion” (objective3). I’ll assume he means that only God-based morality can be objective3, because God-based morality is clearly not objective2 (Craig’s God is a person, a sentient being).
And yet, Craig says that we need God in order to have objective3 morality as if this should be a big deal. But hold on. Even a moral code defined in terms of the preferences of Washoe the chimpanzee is objective3. So not only is Bill’s claim that only God-based morality can be objective3 false (because Harris’ moral theory is also objective3), but also it’s trivially easy to come up with a moral theory that is ‘objective’ in Craig’s (apparent) sense of the term (that is, objective3).8
Moreover, Harris’ theory of morality is objective in a ‘stronger’ sense than Craig’s theory of morality is. Harris’ theory is objective3 and objective2, while Craig’s theory is merely objective3. Whether he’s doing it consciously or not, I wonder if Craig is using the word ‘objective’ to try to sneak in connotations that don’t actually apply to his claims once you pay attention to what Craig actually means by the word ‘objective’. If Craig told his audience that we need God for morality to be ‘objective’ in the same sense that morality defined in terms of the preferences of a chimpanzee is ‘objective’, would this still still have his desired effect on his audience? I doubt it.
Once you’ve stipulated your use of ‘objective’ and ‘subjective’, it is often trivial to determine whether a given moral reductionism is ‘objective’ or ‘subjective’. But what of it? What force should those words carry after you’ve tabooed them? Be careful not to sneak in connotations that don’t belong.
Relative vs. Absolute Morality
Is morality relative or absolute? Again, it depends which moral reductionism you have in mind, and what you mean by ‘relative’ and ‘absolute’. Again, we must be careful about sneaking in connotations.
Moore’s Open Question Argument
“He’s an unmarried man, but is he a bachelor?” This is a ‘closed’ question. The answer is obviously “Yes.”
In contrast, said G.E. Moore, all questions of the type “Such and such is X, but is it good?” are open questions. It feels like you can always ask, “Yes, but is it good?” In this way, Moore resists the identification of ‘morally good’ with any set of natural facts. This is Moore’s Open Question Argument. Because some moral reductionisms do identify ‘good’ or ‘right’ with a particular X, those reductionisms had better have an answer to Moore.
The Yudkowskian response is to point out that when cognitivists use the term ‘good’, their intuitive notion of ‘good’ is captured by a massive logical function that can’t be expressed in simple statements like “maximize pleasure” or “act only in accordance with maxims you could wish to be a universal law without contradiction.” Even if you think everything you want (or rather, want to want) can be realized by (say) maximizing the well-being of conscious creatures, you’re wrong. Your values are more complex than that, and we can’t see the structure of our values. That is why it feels like an open question remains no matter which simplistic identification of “Good = X” you choose.
The problem is not that there is no way to identify ‘good’ or ‘right’ (as used intuitively, without tabooing) with a certain X. The problem is that X is huge and complicated and we don’t (yet) have access to its structure.
But that’s the response to Moore after righting a wrong question—that is, when doing empathic metaethics. When doing mere pluralistic moral reductionism, Moore’s argument doesn’t apply. If we taboo and reduce, then the question of ”...but is it good?” is out of place. The reply is: “Yes it is, because I just told you that’s what I mean to communicate when I use the word-tool ‘good’ for this discussion. I’m not here to debate definitions; I’m here to get something done.”9
The Is-Ought Gap
(This section rewritten for clarity.)
Many claim that you cannot infer an ‘ought’ statement from a series of ‘is’ statements. The objection comes from Hume, who said he was surprised whenever an argument made of is and is not propositions suddenly shifted to an ought or ought not claim, without explanation.10
The solution is to make explicit the bridge from ‘ought’ statements to ‘is’ statements.
Perhaps the arguer means something non-natural by ‘ought’, such as ‘commanded by God’ or ‘in accord with irreducible, non-natural facts about goodness’ (see Moore). If so, I would reject that premise of the argument, because I’m a reductionist. At this point, our discussion might need to shift to a debate over the merits of reductionism.
Or perhaps by ‘you ought to X’ the arguer means something fully natural, such as:
“X is obligatory (by deontic logic) if you assume axiomatic imperatives Y and Z.”
Or: “X tends to maximizes reward signals in agents exhibiting multiple-drafts consciousness” (or, as Sam Harris more broadly puts it, “X tends to maximize well-being in conscious creatures”).
Or: “X is what a Bayes-rational and Hubble-volume-omniscient agent would do if it was motivated to maximize the amount of non-moral goodness from a view in which the interests of all potentially affected individuals were counted equally, where ‘non-moral goodness’ refers to what an agent would want if it were he to contemplate its present situation from a standpoint fully and vividly informed about itself and its circumstances, and entirely free of cognitive error or lapses of instrumental rationality” (see Railton’s metaethics).
Or: “X maximizes the complicated function that can be computed by extrapolating (in a particular way) the motivations encoded by my brain” (see CEV).
Or: “[insert here whatever statement, if believed, would motivate one to do X]” (see Will Sawin).
Or, the speaker may have in mind a common ought-reductionism known as the hypothetical imperative. This is an ought of the kind: “If you desire to lose weight, then you ought to consume fewer calories than your burn.” (But usually, people leave off the implied if statement, and simply say “You should eat less and exercise more.”)
A hypothetical imperative (as some use it) can be translated from ‘ought’ to ‘is’ in a straightforward way: “If you desire to lose weight, then you ought to consume fewer calories than you burn” translates to the claim “If you consume fewer calories than you burn, then you will (or are, ceteris paribus, more likely to) fulfill your desire to lose weight.”11
Or, the speaker may be using ‘ought’ to communicate something only about other symbols (example: Bayes’ Rule), leaving the bridge from ‘ought’ to ‘is’ to be built when the logical function represented by his use of ‘ought’ is plugged into a theory that refers to the world.
But one must not fall into the trap of thinking that a definition you’ve stipulated (aloud or in your head) for ‘ought’ must match up to your intended meaning of ‘ought’ (to which you don’t have introspective access). In fact, I suspect it never does, which is why the conceptual analysis of ‘ought’ language can go in circles for centuries, and why any stipulated meaning of ‘ought’ is a fake utility function. To see clearly to our intuitive concept of ought, we’ll have to try empathic metaethics (see below).
But whatever our intended meaning of ‘ought’ is, the same reasoning applies. Either our intended meaning of ‘ought’ refers (eventually) to the world of math and physics (in which case the is-ought gap is bridged), or else it doesn’t (in which case it fails to refer).12
Moral realism vs. Anti-realism
So, does all this mean that we can embrace moral realism, or does it doom us to moral anti-realism? Again, it depends on what you mean by ‘realism’ and ‘anti-realism’.
In a sense, pluralistic moral reductionism can be considered a robust form of moral ‘realism’, in the same way that pluralistic sound reductionism is a robust form of sound realism. “Yes, there really is sound, and we can locate it in reality — either as vibrations in the air or as mental auditory experiences, however you are using the term.” In the same way: “Yes, there really is morality, and we can locate it in reality — either as a set of facts about the well-being of conscious creatures, or as a set of facts about what an ideally rational and perfectly informed agent would prefer, or as some other set of natural facts.”
But in another sense, pluralistic moral reductionism is ‘anti-realist’. It suggests that there is no One True Theory of Morality. (We use moral terms in a variety of ways, and some of those ways refer to different sets of natural facts.) And as a reductionist approach to morality, it might also leave no room for moral theories which say there are universally binding moral rules for which the universe (e.g. via a God) will hold us accountable.
What matters are the facts, not whether labels like ‘realism’ or ‘anti-realism’ apply to ‘morality’.
Toward Empathic Metaethics
But pluralistic moral reductionism satisfies only a would-be austere metaethicist, not an empathic metaethicist.
Recall that when Alex asks how she can do what is right, the Austere Metaethicist replies:
Tell me what you mean by ‘right’, and I will tell you what is the right thing to do. If by ‘right’ you mean X, then Y is the right thing to do. If by ‘right’ you mean P, then Z is the right thing to do. But if you can’t tell me what you mean by ‘right’, then you have failed to ask a coherent question, and no one can answer an incoherent question.
Alex may reply to the Austere Metaethicist:
Okay, I’m not sure exactly what I mean by ‘right’. So how do I do what is right if I’m not sure what I mean by ‘right’?
The Austere Metaethicist refuses to answer this question. The Empathic Metaethicist, however, is willing to go the extra mile. He says to Alex:
You may not know what you mean by ‘right.’ But let’s not stop there. Here, let me come alongside you and help decode the cognitive algorithms that generated your question in the first place, and then we’ll be able to answer your question. Then we can tell you what the right thing to do is.
This may seem like too much work. Would we be motivated to decode the cognitive algorithms producing Albert and Barry’s use of the word ‘sound’? Would we try to solve ‘empathic meta-acoustics’? Probably not. We can simply taboo and reduce ‘sound’ and then get some work done.
But moral terms and value terms are about what we want. And unfortunately, we often don’t know what we want. As such, we’re unlikely to get what we really want if the world is re-engineered in accordance with our current best guess as to what we want. That’s why we need to decode the cognitive algorithms that generate our questions about value and morality.
So how can the Empathic Metaethicist answer Alex’s question? We don’t know the details yet. For example, we don’t have a completed cognitive neuroscience. But we have some ideas, and we know of some open problems that may admit of progress once more people understand them. In the next few posts, we’ll take our first look at empathic metaethics.13
Previous post: Conceptual Analysis and Moral Theory
Notes
1 Some have objected that the conceptual analysis argued against in Conceptual Analysis and Moral Theory is not just a battle over definitions. But a definition is “the formal statement of the meaning or significance of a word, phrase, etc.”, and a conceptual analysis is (usually) a “formal statement of the meaning or significance of a word, phrase, etc.” in terms of necessary and sufficient conditions. The goal of a conceptual analysis is to arrive at a definition for a term that captures our intuitions about its meaning. The process is to bash our intuitions against others’ intuitions until we converge upon a set of necessary and sufficient conditions that captures them all. But consider Barry and Albert’s debate over the definition of ‘sound’. Why think Albert and Barry have the same concept in mind? Words mean slightly different things in different cultures, subcultures, and small communities. We develop different intuitions about their meaning based on divergent life experiences. Our intuitions differ from each other’s due to the specifics of unconscious associative learning and attribution substitution heuristics. What is the point of bashing our intuitions about the meaning of terms against each other for thousands of pages, in the hopes that we’ll converge on a precise set of necessary and sufficient conditions? Even if we can get Albert and Barry to agree, what happens when Susan wants to use the same term, but has slightly differing intuitions about its meaning? And, let’s say we arrive at a messy set of 6 necessary and sufficient conditions for the intuitive meaning of the term. Is that going to be as useful for communication as one we consciously chose because it carved-up thingspace well? I doubt it. The IAU’s definition of ‘planet’ is more useful than the folk-intuitions definition of ‘planet’. Folk intuitions about ‘planet’ evolved over thousands of years and different people have different intuitions which may not always converge. In 2006, the IAU used modern astronomical knowledge to carve up thingspace in a more useful and informed way than our intuitions do.
A passage from Bertrand Russell (1953) is appropriate. Russell said that many philosophers reminded him of
the shopkeeper of whom I once asked the shortest way to Winchester. He called to a man in the back premises:
“Gentleman wants to know the shortest way to Winchester.”
“Winchester?” an unseen voice replied.
“Aye.”
“Way to Winchester?”
“Aye.”
“Shortest way?”
“Aye.”
“Dunno.”
He wanted to get the nature of the question clear, but took no interest in answering it. This is exactly what modern philosophy does for the earnest seeker after truth. Is it surprising that young people turn to other studies?
2 Compare also to the biologist’s ‘species concept pluralism’ and the philosopher’s ‘art concept pluralism.’ See Uidhir & Magnus (2011). Also see ‘causal pluralism’ (Godfrey-Smith, 2009; Cartwright, 2007), ‘theory concept pluralism’ (Magnus, 2009) and, especially, ‘metaethical contextualism’ (Bjornsson & Finlay, 2010) or ‘metaethical pluralism’ or ‘metaethical ambivalence’ (Joyce, 2011). Joyce quotes Lewis (1989), who wrote that some concepts of value refer to things that really exist, and some concepts don’t, and what you make of this situation is largely a matter of temperament:
What to make of the situation is mainly a matter of temperament. You can bang the drum about how philosophy has uncovered a terrible secret: there are no values! … Or you can think it better for public safety to keep quiet and hope people will go on as before. Or you can declare that there are no values, but that nevertheless it is legitimate—and not just expedient—for us to carry on with value-talk, since we can make it all go smoothly if we just give the name of value to claimants that don’t quite deserve it… Or you can think it an empty question whether there are values: say what you please, speak strictly or loosely. When it comes to deserving a name, there’s better and worse but who’s to say how good is good enough? Or you can think it clear that the imperfect deservers of the name are good enough, but only just, and say that although there are values we are still terribly wrong about them. Or you can calmly say that value (like simultaneity) is not quite as some of us sometimes thought. Myself, I prefer the calm and conservative responses. But as far as the analysis of value goes, they’re all much of a muchness.
Joyce concludes that, for example, the moral naturalist and the moral error theorist may agree with each other (when adopting each other’s own language):
[Metaethical ambivalence] begins with a kind of metametaethical enlightenment. The moral naturalist espouses moral naturalism, but this espousal reflects a mature decision, by which I mean that the moral naturalist doesn’t claim to have latched on to an incontrovertiblerealm of moral facts of which the skeptic is foolishly ignorant, but rather acknowledges that this moral naturalism has been achieved only via a non-mandatory piece of conceptual precisification. Likewise, the moral skeptic champions moral skepticism, but this too is a sophisticated verdict: not the simple declaration that there are no moral values and that the naturalist is gullibly uncritical, but rather a decision that recognizes that this skepticism has been earned only by making certain non-obligatory but permissible conceptual clarifications.
...The enlightened moral naturalist doesn’t merely (grudgingly) admit that the skeptic is warranted in his or her views, but is able to adopt the skeptical position in order to gain the insights that come from recognizing that we live in a world without values. And the enlightened moral skeptic goes beyond (grudgingly) conceding that moral naturalism is reasonable, but is capable of assuming that perspective in order to gain whatever benefits come from enjoying epistemic access to a realm of moral facts.
3 Miller (2003), p. 3.
4 I changed the example moral judgment from “murder is wrong” to “stealing is wrong” because the former invites confusion. ‘Murder’ often means wrongful killing.
5 Also see Jacobs (2002), starting on p. 2.
6 The first premise of one of his favorite arguments for God’s existence is “If God does not exist, objective moral values and duties do not exist.”
7 Craig (2010), p. 11.
8 It’s also possible that Craig intended a different sense of objective than the ones explicitly given in his article. Perhaps he meant objective4: “morality is objective4 if it is not grounded in the opinion of non-divine persons.”
9 Also see Moral Reductionism and Moore’s Open Question Argument.
10 Hume (1739), p. 469. The famous paragraph is:
In every system of morality, which I have hitherto met with, I have always remarked, that the author proceeds for some time in the ordinary ways of reasoning, and establishes the being of a God, or makes observations concerning human affairs; when all of a sudden I am surprised to find, that instead of the usual copulations of propositions, is, and is not, I meet with no proposition that is not connected with an ought, or an ought not. This change is imperceptible; but is however, of the last consequence. For as this ought, or ought not, expresses some new relation or affirmation, ’tis necessary that it should be observed and explained; and at the same time that a reason should be given; for what seems altogether inconceivable, how this new relation can be a deduction from others, which are entirely different from it.
11 For more on reducing certain kinds of normative statements, see Finlay (2010).
12 Assuming reductionism is true. If reductionism is false, then of course there are problems for pluralistic moral reductionism as a theory of austere (but not empathic) metaethics. The clarifications in the last three paragraphs of this section are due to discussions with Wei Dai and Vladimir Nesov.
13 My thanks to Steve Rayhawk and Will Newsome for their feedback on early drafts of this post.
References
Bjornsson & Finlay (2010). Metaethical contextualism defended. Ethics, 121: 7-36.
Craig (2010). Five Arguments for God. The Gospel Coalition.
Cartwright (2007).Hunting Causes and Using Them: Approaches in Philosophy and Economics. Cambridge University Press.
Godfrey-Smith (2009). Causal pluralism. In Beebee, Hitchcock, & Menzies (eds.), The Oxford Handbook of Causation (pp. 326-337). Oxford University Press.
Hume (1739). A Treatise on Human Nature. John Noon.
Finlay (2010). Normativity, Necessity and Tense: A Recipe for Homebaked Normativity. In Shafer-Landau (ed.), Oxford Studies in Metaethics 5 (pp. 57-85). Oxford University Press.
Jacobs (2002). Dimensions of Moral Theory. Wiley-Blackwell.
Joyce (2011).Metaethical pluralism: How both moral naturalism and moral skepticism may be permissible positions. In Nuccetelli & Seay (eds.), Ethical Naturalism: Current Debates. Cambridge University Press.
Lewis (1989). Dispositional theories of value. Part II. Proceedings of the Aristotelian Society, supplementary vol. 63: 113-137.
Magnus (2009). What species can teach us about theory.
Miller (2003). An Introduction to Contemporary Metaethics. Polity.
Russell (1953). The cult of common usage. British Journal for the Philosophy of Science, 12: 305-306.
Uidhir & Magnus (2011). Art concept pluralism. Metaphilosophy, 42: 83-97.
- A Crash Course in the Neuroscience of Human Motivation by 19 Aug 2011 21:15 UTC; 203 points) (
- Conceptual Analysis and Moral Theory by 16 May 2011 6:28 UTC; 94 points) (
- Concepts Don’t Work That Way by 28 Sep 2011 2:01 UTC; 88 points) (
- Population Ethics Without Axiology: A Framework by 2 Aug 2022 15:59 UTC; 81 points) (EA Forum;
- What Is Moral Realism? by 22 May 2018 15:49 UTC; 72 points) (EA Forum;
- Why Realists and Anti-Realists Disagree by 5 Jun 2020 7:51 UTC; 61 points) (EA Forum;
- The Life-Goals Framework: How I Reason About Morality as an Anti-Realist by 3 Feb 2022 11:40 UTC; 48 points) (EA Forum;
- Against Irreducible Normativity by 9 Jun 2020 14:38 UTC; 48 points) (EA Forum;
- Realism and Rationality by 16 Sep 2019 3:09 UTC; 45 points) (
- Rudimentary Categorization of Less Wrong Topics by 5 Sep 2015 7:32 UTC; 39 points) (
- End-Relational Theory of Meta-ethics: A Dialogue by 28 Jun 2016 20:11 UTC; 38 points) (EA Forum;
- Quick thoughts on empathic metaethics by 12 Dec 2017 21:46 UTC; 29 points) (
- Moral uncertainty vs related concepts by 11 Jan 2020 10:03 UTC; 26 points) (
- 26 Nov 2019 7:50 UTC; 23 points) 's comment on Realism and Rationality by (
- My intentions for my metaethics sequence by 30 Aug 2011 16:52 UTC; 22 points) (
- 3 Nov 2012 22:19 UTC; 17 points) 's comment on In Defense of Moral Investigation by (
- Scientist vs. philosopher on conceptual analysis by 20 Sep 2011 15:10 UTC; 17 points) (
- Meanings of Mathematical Truths by 5 Jun 2011 22:59 UTC; 13 points) (
- A Defense of Naive Metaethics by 9 Jun 2011 17:46 UTC; 13 points) (
- 13 Feb 2013 7:58 UTC; 13 points) 's comment on Questions for Moral Realists by (
- Metaethics: Where I’m Headed by 22 Nov 2011 8:15 UTC; 13 points) (
- Questions for Moral Realists by 13 Feb 2013 5:44 UTC; 8 points) (
- 1 Sep 2022 11:21 UTC; 7 points) 's comment on On the Philosophical Foundations of EA by (EA Forum;
- 27 Jun 2011 23:18 UTC; 6 points) 's comment on Learning how to explain things by (
- 6 Apr 2012 23:34 UTC; 6 points) 's comment on Complexity of Value ≠ Complexity of Outcome by (
- 29 Jun 2011 18:38 UTC; 6 points) 's comment on A Defense of Naive Metaethics by (
- 7 Jul 2011 5:28 UTC; 6 points) 's comment on A Defense of Naive Metaethics by (
- 26 Sep 2012 14:39 UTC; 6 points) 's comment on [Poll] Less Wrong and Mainstream Philosophy: How Different are We? by (
- 26 Nov 2019 7:52 UTC; 5 points) 's comment on I’m Buck Shlegeris, I do research and outreach at MIRI, AMA by (EA Forum;
- 1 Jun 2011 1:08 UTC; 5 points) 's comment on Pluralistic Moral Reductionism by (
- 1 Jun 2011 2:05 UTC; 5 points) 's comment on Pluralistic Moral Reductionism by (
- 22 May 2018 22:10 UTC; 4 points) 's comment on What Is Moral Realism? by (EA Forum;
- Amending the “General Pupose Intelligence: Arguing the Orthogonality Thesis” by 13 Mar 2013 23:21 UTC; 4 points) (
- 30 Aug 2011 22:56 UTC; 4 points) 's comment on My intentions for my metaethics sequence by (
- 19 Nov 2014 20:15 UTC; 3 points) 's comment on Conceptual Analysis and Moral Theory by (
- 24 Nov 2011 8:29 UTC; 3 points) 's comment on Metaethics: Where I’m Headed by (
- 11 Dec 2012 17:59 UTC; 3 points) 's comment on By Which It May Be Judged by (
- ‘Is’ and ‘Ought’ and Rationality by 5 Jul 2011 3:53 UTC; 2 points) (
- 1 Sep 2013 20:05 UTC; 2 points) 's comment on P/S/A—Sam Harris offering money for a little good philosophy by (
- Questions for Moral Realists by 13 Feb 2013 5:42 UTC; 2 points) (
- 8 Jul 2017 22:22 UTC; 1 point) 's comment on Hi, I’m Luke Muehlhauser. AMA about Open Philanthropy’s new report on consciousness and moral patienthood by (EA Forum;
- 16 Aug 2011 19:44 UTC; 1 point) 's comment on Are Deontological Moral Judgments Rationalizations? by (
- 3 Nov 2021 23:03 UTC; 1 point) 's comment on 2020 PhilPapers Survey Results by (
- 21 Jun 2011 19:13 UTC; 1 point) 's comment on A Defense of Naive Metaethics by (
- 12 Sep 2011 4:02 UTC; 0 points) 's comment on A Sketch of an Anti-Realist Metaethics by (
- 7 Jul 2011 23:09 UTC; 0 points) 's comment on A Defense of Naive Metaethics by (
- 8 Jul 2011 15:05 UTC; 0 points) 's comment on A Defense of Naive Metaethics by (
- 8 Jul 2011 15:33 UTC; 0 points) 's comment on A Defense of Naive Metaethics by (
- 1 Mar 2013 16:02 UTC; 0 points) 's comment on Decision Theory FAQ by (
- 21 Nov 2013 17:45 UTC; -4 points) 's comment on No Universally Compelling Arguments in Math or Science by (
- Natural Rights as Impediment to Artificial Intelligence by 31 Mar 2014 23:23 UTC; -24 points) (
Solid and unsurprising.
Thanks, this is exactly the feedback I was hoping to receive. :)
Basically, I want this post and the last one to be where Less Wrongers can send people whenever they appear confused about standard philosophical debates in moral theory: “Wait, stop. Go read lukeprog’s article on this and then let me know if you still think the same thing.”
I feel like your austere meta-ethicist is mostly missing the point. It’s utterly routine for different people to have conflicting beliefs about whether a given act is moral*. And often they can have a useful discussion, at the end of which one or both participants change their beliefs. These conversations can happen without the participants changing their definitions of words like ‘moral’, and often without them having a clear definition at all.
[This is my first LW comment—if I do something wrong, please bear with me]
This suggests that precise definitions or agreement about definitions isn’t all that critical. But it’s sometimes useful to be able to reason from stipulated and mutually agreed definitions, in which case meta-ethical speculation and reasoning is doing useful work if it offers a menu of crisp, useful, definitions that can be used in discussion of specific moral claims. Relatedly, it’s also doing useful work by offering a set of definitions that help people conceptualize and articulate their personal feelings about morality, even absent a concrete first-order question.
And part of what goes into picking definitions is to understand their consequences. A philosopher is doing useful work for me if he shows me that a tempting-sounding definition of ‘morality’ doesn’t pick out the set of things I want it to pick out, or that some other definition turns out not to refer to any clear set at all.
Many mathematical entities have multiple logically equivalent definitions, that are of different utility in different contexts. (E.g., sometimes I want to think about a circle as a locus of points, and sometimes as the solution set to an equation.) In the real world, something similar happens.
When I discuss, say, abortion, with somebody, probably there are multiple working definitions of ‘moral’ that could be mutually agreed upon for the purpose of the conversation, and the underlying dispute would still be nontrivial and intelligible. But some definitions might be more directly applicable to the discussion—and philosophical reasoning might be helpful in figuring out what the consequences of various definitions are. For instance, a non-cognitive strikes me intuitively as less likely to be useful—but I’d be open to an argument showing how it could be useful in a debate.
Probably a great deal of academic writing on meta-ethics is low value. But that’s true of most writing on most topics and doesn’t show that the topic is pointless. (With academics being major offenders, but not the only offenders.)
*I’m thinking of the individual personal changes in belief that went along with increased opposition to official racism in America over the course of the 20th century. Or opposition to slavery in the 19th.
Welcome to Less Wrong!
Is there a part of your comment that you suspect I disagree with? Or, is there a sentence in my post with which you disagree?
Having had time to mull over—I think here’s something about your post that bothers me. I don’t think it’s possible to pinpoint a single sentence, but here are two things that don’t quite satisfy me.
1) Neither your austere or empathetic meta-ethicists seem to be telling me anything I wanted to hear. What I want is a “linguistic meta-ethicist”, who will tell me what other competent speakers of English mean when they use “moral” and suchlike terms. I understand that different people mean different things, and I’m fine with an answer which comes in several parts, and with notes about which speakers are primarily using which of those possible definitions.
What I don’t want is a brain scan from each person I talk to—I want an explanation that’s short and accessible enough to be useful in conversations. Conventional ethics and meta-ethics has given a bunch of useful definitions. Saying “well, it depends” seems unnecessarily cautious; saying “let’s decode your brain” seems excessive for practical purposes.
2) Most of the conversations I’m in that involve terms like “moral” would be only slightly advanced by having explicit definitions—and often the straightforward terms to use instead of “moral” are very nearly as contentious or nebulous. In your own examples, you have your participants talk about “well-being” and “non-moral goodness.” I don’t think that’s a significant step forward. That’s just hiding morality inside the notion of “a good life”—which is a sensible thing to say, but people have been saying it since Plato, and it’s an approach that has problems of its own.
By the way, I do understand that I may not have been your target audience, and that the whole series of posts has been carefully phrased and well organized, and I appreciate that.
I would think that the Hypothetical Imperatives are useful there. You can thus break down your own opinions into material of the form:
“If the set X of imperative premises holds, and the set Y of factual premises holds, then logic Z dictates that further actions W are imperative.
“I hold X already, and I can convince logic Z of the factual truth of Y, thus I believe W to be imperative.”
Even all those complete bastards who disagree with your X can thus come to an agreement with you about the hypothetical as a whole, provided they are epistemically rational. Having isolated the area of disagreement to X, Y, or Z, you can then proceed to argue about it.
asr,
Your linguistic metaethicist sounds like the standard philosopher doing conceptual analysis. Did you see my post on ‘Conceptual Analysis and Moral Theory’?
I think conversations using moral terms would be greatly advanced by first defining the terms of the debate, as Aristotle suggested. Also, the reason ‘well-being’ or ‘non-moral goodness’ are not unpacked is because I was giving brief examples. You’ll notice the austere metaethicist said things like “assuming we have the same reduction of well-being in mind...” I just don’t have the space to offer such reductions in what is already a long post.
I would find it helpful—and I think several of the other posters here would as well—to see one reduction on some nontrivial question carried far enough for us to see that the process can be made to work. If I understand right, your approach requires that speakers, or at least many speakers much of the time, can reduce from disputed, loaded, moral terms to reasonably well-defined and fact-based terminology. That’s the point I’d most like to see you spend your space budget on in future posts.
Definitions are good. Precise definitions are usually better than loose definitions. But I suspect that in this context, loose definitions are basically good enough and that there isn’t a lot of value to be extracted by increased precision there. I would like evidence that improving our definitions is a fruitful place to spend effort.
I did read your post on conceptual analysis. I just re-read it. And I’m not convinced that the practice of conceptual analysis is any more broken than most of what people get paid to do in the humanities and social sciences . My sense is that the standard textbook definitions are basically fine, and that the ongoing work in the field is mostly just people trying to get tenure and show off their cleverness.
I don’t see that there’s anything terribly wrong with the practice of conceptual analysis—so long as we don’t mistake an approximate and tentative linguistic exercise for access to any sort of deep truth.
I don’t think many speakers actually have an explicit ought-reduction in mind when they make ought claims. Perhaps most speakers actually have little idea what they mean when they use ought terms. For these people, emotivism may roughly describe speech acts involving oughts.
Rather, I’m imagining a scenario where person A asks what they ought to do, and person B has to clarify the meaning of A’s question before B can give an answer. At this point, A is probably forced to clarify the meaning of their ought terms more thoroughly than they have previously done. But if they can’t do so, then they haven’t asked a meaningful question, and B can’t answer the question as given.
Why? What I’ve been saying the whole time is that improving our definitions isn’t worth as much effort as philosophers are expending on it.
On this, we agree. That’s why conceptual analysis isn’t very valuable, along with “most of what people get paid to do in the humanities and social sciences.” (Well, depending on where you draw the boundary around the term ‘social sciences.’)
Do you see something wrong with the way Barry and Albert were arguing about the meaning of ‘sound’ in Conceptual Analysis and Moral Theory? I’m especially thinking of the part about microphones and aliens.
I agree that emotivism is an accurate description, much of the time, for what people mean when they make value judgments. I would also agree that most people don’t have a specific or precise definition in mind. But emotivism isn’t the only description and for practical purposes it’s often not the most useful. Among other things, we have to specify which emotion we are talking about. Not all disgust is moral disgust.
Value judgments show up routinely in law and in daily life. It would be an enormous, difficult, and probably low-value task to rewrite our legal code to avoid terms like “good cause”, “unjust enrichment”, “unconscionable contract”, and the like. Given that we’re stuck with moral language, it’s a useful project to pull out some definitions to help focus discourse slightly. But we aren’t going to be able to eliminate them. “Morality” and its cousins are too expensive to taboo.
We want law and social standards to be somewhat loosely defined, to avoid unscrupulous actors trying to worm their way through loopholes. We don’t want to be overly precise and narrow in our definitions—we want to leverage the judgement of judges and juries. But conversely, we do want to give them guidance about what we mean by those words. And precedent supplies one sort of guidance, and some definitions give them an additional sort of guidance.
I suspect it would be quite hard to pick out precisely what we as a society mean when we use those terms in the legal code—and very hard to reduce them to any sort of concrete physical description that would still be human-intelligible. I would be interested to see a counterexample if you can supply one easily.
I have the sense that trying to talk about human judgement and society without moral language would be about like trying to discuss computer science purely in terms of the hardware—possible, but unnecessarily cumbersome.
One of the common pathologies of the academy is that somebody comes up with a bright idea or a powerful intellectual tool. Researchers then spend several years applying that tool to increasingly diverse contexts, often where the marginal return from the tool is near-zero. Just because conceptual analysis is being over-used doesn’t mean that it is always useless! The first few uses of it may indeed have been fairly high-value in aiding us in communicating. The fact that the tool is then overused isn’t a reason to ignore it.
Endless wrangles about definitions, I think are necessarily low-value. Working out a few useful definitions or explanations for a common term can be valuable, though—particularly if we are going to apply those terms in a quasi-formal setting, like law.
It may be routine in the sense that it often happens, but not routine in the sense that this is a reliable approach to settling moral differences. Often such disputes are not settled despite extensive discussions and no obvious disagreement about other kinds of facts.
This can be explained if individuals are basing their judgments off differing sets of values that partially overlap. Even if both participants are naively assuming their own set of values is the set of moral values, the fact of overlapping will sometimes mean non-moral considerations which are significant to one’s values will also be significant for the other’s values. Other times, this won’t be the case.
For example, many pro-lifers naively assume that everyone places very high value on all human organisms, so they spend a lot of time arguing that an embryo or fetus is a distinct human organism. Anyone who is undecided or pro-choice who shares this value but wasn’t aware of the biological evidence that unborn humans are distinct organisms from their mothers may be swayed by such considerations.
On the other hand, many pro-choicers simply do not place equally high value on all human organisms, without counting other properties like sentience. Or — following Judith Jarvis Thomson in “A Defense of Abortion” — they may place equally high value on all human organisms, but place even greater value on the sort of bodily autonomy denied by laws against abortion.
Morality as the expression of pluralistic value sets (and the hypothetical imperatives which go along with them) is a very neat explanation of the pattern we see of agreement, disagreement, and partially successful deliberation.
I agree with all the claims you’re making about morality and about moral discussion. But I don’t quite see where any of this is giving me any new insights or tools. Sure, people have different but often overlapping values. I knew that. I think most adults who ever have conversations about morality know that. And we know that without worrying too much about the definition of morality and related words.
But I think everything you’ve said is also true about personal taste in non moral questions. I and my friends have different but overlapping taste in music, because we have distinct but overlapping set of desiderata for what we listen to. And sometimes, people get convinced to like something they previously didn’t. I want a meta-ethics that gives me some comparative advantage in dealing with moral problems, as compared to other sorts of disagreements. I had assumed that lukeprog was trying to say something specifically about morality, not just give a general and informal account of human motivation, values, and preferences.
Thus far, this sequence feels like a lot of buildup and groundwork that is true but mostly not in much dispute and mostly doesn’t seem to help me accomplish anything. Perhaps my previous comment should just have been a gentle nudge to lukeprog to get to the point.
This may be a case where not getting it wrong is the main point, even if getting it right is a let down.
My own view is quite similar to Luke’s and I find it useful when I hear a moral clam to try sorting out how much of the claim is value-expression and how much is about what needs to be done to promote values. Even if you don’t agree about values, it still helps to figure out what someone else’s fundamental values are and argue that what they’re advocating is out of line with their own values. People tend to be mistaken about how to fulfill their own values more than they are about how to fulfill their own taste in music.
Yes.
That is why I can interrogate what somebody means by ‘ought’ and then often show that by their own definition of ought, what they thought they ‘ought’ to do is not what they ‘ought’ to do.
Do you know of anything better?
OTOH, the problem remains that people act on their values, and that one persons actions can affect another person. Pluralistic morality is terrible at translating into a uniform set of rules that all are beholden to.
Why is that the test of a metaethical theory rather than the theory which best explains moral discourse? Categorical imperatives — if that’s what you’re referring to — are one answer to the best explanation of moral discourse, but then we’re stuck showing how categorical imperatives can hold...or accepting error theory.
Perhaps ‘referring to categorical imperatives’ is not the only or even the best explanation of moral discourse. See “The Error in the Error Theory” by Stephen Finlay.
Because there is a practical aspect to ethics. Moral discourse involves the idea that people should do the obligatory and refrain from the forbidden. -- irrespective of who they are. That needs explaining as well.
Moral discourse is about what to do, but it doesn’t seem to (at least always) be about what everyone must do for no prior reason.
Uh-huh. Is that an issue of commission rather than omission? Are people not obligated to refrain from theft murder and rape , their inclinations notwithstanding?
If by ‘obligated’ you mean it’s demanded by those who fear being the targets of those actions, yes. Or if you mean exercising restraint may be practically necessary to comply with certain values those actions thwart, yes. Or if you mean doing those things is likely to result in legal penalties, that’s often the case.
But if you mean it’s some simple fact that we’re morally obligated to restrain ourselves from doing certain things, no. Or at least I don’t see how that could even possibly be the case, and I already have a theory that explains why people might mistakenly think such a thing is the case (they mistake their own values for facts woven into the universe, so hypothetical imperatives look like categorical imperatives to them).
The ‘commission’ vs. ‘omission’ thing is often a matter of wording. Rape can be viewed as omitting to get proper permission, particularly when we’re talking about drugging, etc.
Well, I have a theory about how it could be the case. Objective morality doesn’;t have to be a fact-like thing that is paradoxically indetectable. It could be based on the other source of objectivity: logic and reason. It’s an analytical truth that you shouldn’t do to others what you wouldn’t want done to yourself. You are obliged to be moral so long as you can reason morally in the sense that you will be held responsible.
I’m skeptical that this statement is true, let alone an analytic truth. Different people have different desires. I take the Golden Rule to be a valuable heuristic, but no more than that.
What is your reason for believing that it is true as an absolute rule?
Just to clarify where you stand on norms: Would you say a person is obligated by facts woven into the universe to believe that 68 + 57 = 125 ? (ie, are we obligated in this sense to believe anything?)
To stick my own neck out: I am a realist about values. I think there are facts about what we ought to believe and do. I think you have to be, to capture mathematical facts. This step taken, there’s no further commitment required to get ethical facts. Obviously, though, there are epistemic issues associated with the latter which are not associated with the former.
Would it be fair to extrapolate this, and say that individual variation in value sets provides a good explanation of the pattern we see of agreement and disagreement between individuals as regards moral values—and possibly in quite different domains as well (politics, aesthetics, gardening)?
You seem to be suggesting meta-ethics aims merely to give a discriptively adequate characterisation of ethical discourse. If so, would you at least grant that many see (roughly) as its goal to give a general characterisation of moral rightness, that we all ought to strive for it?
Facts as in true statements, or facts as in states-of-affiairs?
Facts in the disappointingly deflationary sense that
It is a fact that P if and only if P (and that’s all there is to say about facthood).
This position is a little underwhelming to any who seek a metaphysically substantive account of what makes things true, but it is a realist stance all the same (no?). If you have strong arguments against this or for an alternative, I’m interested to hear.
No, I wouldn’t say that. It would be a little odd to say anyone who doesn’t hold a belief that 68 + 57 equals 125 is neglecting some cosmic duty. Instead, I would affirm:
In order to hold a mathematically correct belief when considering 68 + 57, we are obligated to believe it equals 125 or some equivalent expression.
(I’m leaving ‘mathematically correct’ vague so different views on the nature of math are accommodated.)
In other words, the obligation relies on a goal. Or we could say normative answers require questions. Sometimes the implied question is so obvious, it seems strange to bother identifying it.
Yes.
I think that’s generally the job of normative ethics, and metaethics is a little more open ended than that. I do grant that many people think the point of ethical philosophy in general is to identify categorical imperatives, not give a pluralistic reduction.
Taking your thoughts out of order,
What I was getting at is that this looks like complete moral relativism -‘right for me’ is the only right there is (since you seem to be implying there is nothing interesting to be said about the process of negotiation which occurs when people’s values differ). I’m understanding that you’re willing to bite this bullet.
I take your point here. I may be conflating ethical and meta-ethical theory. I had in mind theories like Utilitarianism or Kantian ethical theory, which are general accounts of what it is for an action to be good, and do not aim merely to be accurate descriptions of moral discourse (would you agree?). If we’re talking about a defence of, say, non-cognitivism, though, maybe what you say is fair.
This is fair.
This is an interesting proposal, but I’m not sure what it implies. Is it possible for a rational person to strive to believe anything but the truth? Whether in math or anything else, doesn’t a rational person always try to believe what is correct? Or, to put the point another way, isn’t having truth as its goal part of the concept of belief? If so, I suggest this collapses to something like
*When considering 68 + 57, we are obligated to believe it equals 125 or some equivalent expression.
or, more plausibly,
*When considering 68 + 57, we ought to believe it equals 125 or some equivalent expression.
But if this is fair I’m back to wondering where the ought comes from.
While it is relativism, the focus is a bit different from ‘right for me.’ More like ‘this action measures up as right against standard Y’ where this Y is typically something I endorse.
For example, if you and I both consider a practice morally right and we do so because it measures up that way against the standard of improving ‘the well-being of conscious creatures,’ then there’s a bit more going on than it just being right for you and me.
Or if I consider a practice morally right for the above reason, but you consider it morally wrong because it falls afoul of Rawls’ theory of justice, there’s more going on than it just being right for me and wrong for you. It’s more like I’m saying it’s right{Harris standard} and you’re saying it’s wrong{Rawls standard}. (...at least as far as cognitive content is concerned; we would usually also be expressing an expectation that others adhere to the standards we support.)
Of course the above are toy examples, since people’s values don’t tend to line up neatly with the simplifications of philosophers.
It’s not apparent that values differ just because judgments differ, so there’s still a lot of interesting work to find out if disagreements can be explained by differing descriptive beliefs. But, yes, once a disagreement is known to result from a pure difference in values, there isn’t a rational way to resolve it. It’s like Luke’s ‘tree falling’ example; once we know two people are using different definitions of ‘sound,’ the best we can do is make people aware of the difference in their claims.
Yep. While those are interesting standards to consider, it’s pretty clear to me that real world moral discourse is wider and more messy than any one normative theory. We can simply declare a normative theory as the moral standard — plenty of people have! — but the next person whose values are a better match for another normative theory is just going to disagree. On what basis do we find that one normative theory is correct when, descriptively, moral pluralism seems to characterize moral discourse?
If being rational consists in doing what it takes to fulfill one’s goals (I don’t know what the popular definition of ‘rationality’ is on this site), then it is still possible to be rational while holding a false belief, if a false belief helps fulfill one’s goals.
Now typically, false beliefs are unhelpful in this way, but I know at least Sinnott-Armstrong has talked about an ‘instrumentally justified’ belief that can go counter to having a true belief. The example I’ve used before is an Atheist married to a Theist whose goal of having a happy marriage would in fact go better if she could take a belief-altering pill so she would falsely take on her spouse’s belief in God.
Perhaps it comes from the way you view the concept of belief as implying a goal?
At risk of triggering the political mind-killer, I think there are some potentially problematic consequences of this view.
Suppose we don’t have good grounds for keeping one set of moral beliefs over another. Now suppose somebody offers to reward us for changing our views, or punish us for not changing. Should we change our views?
To go from the philosophical to the concrete: There are people in the world who are fanatics who are largely committed to some reading of the Bible/Koran/Little Green Book of Colonel Gaddafi/juche ideology of the Great Leader/whatever. Some of those people have armies and nuclear weapons. They can bring quite a lot of pressure to bear on other individuals to change their views to resemble those of the fanatic.
If rationalism can’t supply powerful reasons to maintain a non-fanatical worldview in the face of pressure to self-modify, that’s an objection to rationalism. Conversely, altering the moral beliefs of fanatics with access to nuclear weapons strikes me as an extremely important practical project. I suspect similar considerations will apply if you consider powerful unfriendly powerful AIs.
This reminds me of that line of Yeats, that “the best lack all conviction, while the worst are full of passionate intensity.” Ideological differences sometimes culminate in wars, and if you want to win those wars, you may need something better than “we have our morals and they have theirs.”
To sharpen the point slightly: There’s an asymmetry between the rationalists and the fanatics, which is that the rationalists are aware that they don’t have a rational justification for their terminal values, but the fanatic does have a [fanatical] justification. Worse, the fanatic has a justification to taboo thinking about the problem, and the rationalist doesn’t.
Just because morality is personal doesn’t make it not real. If you model people as agents with utility functions, the reason not to change is obvious—if you change, you won’t do all the things you value. Non-fanatics can do that the same as fanatics.
The difference comes when you factor in human irrationality. And sure, fanatics might resist where anyone sane would give in. “We will blow up this city unless you renounce the Leader,” something like that. But on the other hand, rational humans might resist techniques that play on human irrationality, where fanatics might even be more susceptible than average. Good cop / bad cop for example.
What about on a national scale, where, say, an evil mastermind threatens to nuke every nation that does not start worshiping the flying spaghetti monster? Well, what a rational society would do is compare benefits and downsides, and worship His Noodliness if it was worth it. Fanatics would get nuked. I fail to see how this is an argument for why we shouldn’t be rational.
And that’s why Strawmansylvania has never won a single battle, I agree. Just because morality is personal doesn’t make it unmomving.
Does this imply that if a rational actor has terminal values that are internally consistent and in principle satisfiable, it would always be irrational for the actor to change those values or allow them to change?
That doesn’t seem right either. Somehow, an individual improving their moral beliefs as they mature, the notional Vicar of Bray, and Pierre Laval are all substantially different cases of people changing their [terminal] beliefs in response to events. There’s something badly wrong with a theory that can’t distinguish those cases.
Also, my apologies if this has been already discussed to death on LW or elsewhere—I spent some time poking and didn’t see anything on this point.
No, but it sets a high standard—If you value, say, the company of your family, then modifying to not want that (and therefore not spend much time with your family) costs as much as if you were kept away from your family by force for the rest of your life. So any threats have to be pretty damn serious, and maybe not even death would work if you know important secrets or do not highly value living without some key values.
I wouldn’t call all of those cases of modifying terminal values. From some quick googling (I didn’t know about the Vicar of Bray), what the Vicar of Bray cared about was being the vicar of Bray. What Pierre Laval cared about was being the head of the government and not being killed, maybe. So they’re maybe not good examples of changing terminal values, as opposed to instrumental ones.
Also “improving their moral beliefs as they mature” is a very odd concept once you think about it. How do you judge whether a moral belief is right to hold correctly without having a correct ultimate belief from the start, to do the judging? It’s really an example of how humans are emphatically not rational agents—we follow a bunch of evolved and cultural rules, which can appear to produce consistent behavior, but really have all these holes and internal conflicts. And things can change suddenly, without the sort of rational deliberation described above.
You could say the same about “improving our standards of scientific inference.” Circular? Perhaps, but it needn’t be a vicious circle. It’s pretty clear that we’ve accomplished it, so it must be possible.
I would cheerfully agree that humans aren’t rational and routinely change their minds about morality for non-rational reasons.
This is one of the things I was trying to get at. Ask when we should change our minds for non-rational reasons, and when we should attempt to change others’ minds using non-rational means.
The same examples I mentioned above work for these questions too.
Here’s what I had in mind with the reference to the Vicar of Bray. Imagine an individual with two terminal values: “Stay alive and employed” and the reigning orthodoxy at the moment. The individual sincerely believes in both, and whenever they start to conflict, changes their beliefs about the orthodoxy. He is quite sincere in advocating for the ruling ideology at each point in time; he really does believe in divine right of kings, just so long as it’s not a dangerous belief to hold.
The beliefs in question are at least potentially terminal moral beliefs. Without delving deep into the history, let’s stipulate for the purpose of the conversation that we’re talking about a rational actor who has a sequence of terminal moral beliefs about what constitutes a just government, and that these beliefs shift with the political climate.
Now for contrast, let’s consider a hypothetical rational but very selfish child. The child’s parents attempt and succeed in changing the child’s values to be less selfish. They do this by the usual parental tactics of punishment and example-setting, not by rational argument. By your social standard and mine, this is an improvement to the child.
Both the vicar and the child are updating their moral beliefs in response to outside pressure, not rational deliberation. The general consensus is that parents are obligated to bring up their children not to be overly self-centered and that reasoning with children is not a sufficient pedagogic technique But conversely that coercive government pressure on religion is ignoble.
Is this simply that you and I think “a change in moral beliefs, brought about by non-reasonable means is good (all else equal), if it significantly improves the beliefs of the subject by my standards”?
I’d agree with that. Maybe with some caveats, but generally yes.
I think the caveats will turn out to matter a lot. One of the things that human moral beliefs do, in practice, is give other humans some reasons to trust you. If I know that you are committed, for non-instrumental reasons, to avoid manipulating* me into changing my values, that gives me reasons to trust you. Conversely, if your moral view is that it’s legitimate to lie to people to make them do what you want, people will trust you less.
Obviously, people have incentives to lie about their true values. I think equally obviously, people are paying attention and looking hard for that sort of hypocrisy.
*This sentence is true for a range of possible expansions of “manipulating”.
My statement was more observational than ideal, though. Sure, a rational agent can be averse to manipulating other people (and humans often are too), because agents can care about whatever they want. But that doesn’t bear very strongly on how the language is used compared to the fact that in real-world usage I see people say things like “improved his morals” by only three standards: consistency, how much society approves, and how much you approve.
I think the worry here is that realizing ‘right’ and ‘wrong’ are relative to values might make us give up our values. Meanwhile, those who aren’t as reflective are able to hold more strongly onto their values.
But let’s look at your deep worry about fanatics with nukes. Does their disregard for life have to also be making some kind of abstract error for you to keep and act on your own strong regard for life?
Almost. What I’m worried about is that acknowledging or defining values to be arbitrary makes us less able to hold onto them and less able to convince others to adopt values that are safer for us. I think it’s nearly tautological that right and wrong are defined in terms of values.
The comment about fanatics with nuclear weapons wasn’t to indicate that that’s a particular nightmare of mine. It isn’t. Rather, that was to get at the point that moral philosophy isn’t simply an armchair exercise conducted amongst would-be rationalists—sometimes having a good theory a matter of life and death.
It’s very tempting, if you are firmly attached to your moral beliefs, and skeptical about your powers of rationality (as you should be!) to react to countervailing opinion by not listening. If you want to preserve the overall values of your society, and are skeptical of others’ powers of rational judgement, it’s tempting to have the heretic burnt at the stake, or the philosopher forced to drink the hemlock.
One of the undercurrents in the history of philosophy has been an effort to explain why a prudent society that doesn’t want to lose its moral footings can still allow dissent, including dissent about important values, that risks changing those values to something not obviously better. Philosophers, unsurprisingly, are drawn to philosophies that explain why they should be allowed to keep having their fun. And I think that’s a real and valuable goal that we shouldn’t lose sight of.
I’m willing to sacrifice a bunch of other theoretical properties to hang on to a moral philosophy that explains why we don’t need heresy trials and why nobody needs to bomb us for being infidels.
I don’t see much difference there. Relativist morality doesn’t have to be selfish (although the reverse is probably true).
OK, but what I want to know is how you react to some person -whose belief system is internally consistent- who has just, say, committed a gratuitous murder. Are you committed to saying that there are no objective grounds to sanction him—there is no sense in which he ought not to have done what he did (assuming his belief system doesn’t inveigh against him offending yours)?
Touche.
Look, what I’m getting at is this. I assume we can agree that
“68 + 57 = 125” is true if and only if 68 + 57 = 125
This being the case, if A, seriously pondering the nature of the compulsion behind mathematical judgement, should ask, “Why ought I to believe that 68 + 57 = 125?”, and B answers, “Because it’s true”, then B is not really saying anything beyond, “Because it does”. B does not answer A’s question.
If the substantive answer is something along the lines that it is a mathematical fact, then I am interested to know how you conceive of mathematical facts, and whether there mightn’t be moral facts of generally the same ilk (or if not, why). But if you want somehow to reduce this to subjective goals, then it looks to me that mathematics falls by the wayside—you’ll surely allow this looks pretty dubious at least superficially.
There’s an ambiguity here. A standard can make objective judgments, without the selection of that standard being objective. Like meter measurements.
Such a person would be objectively afoul of a standard against randomly killing people. But let’s say he acted according to a standard which doesn’t care about that; we wouldn’t be able to tell him he did something wrong by that other standard. Nor could we tell him he did something wrong according to the one, correct standard (since there isn’t one).
But we can tell him he did something wrong by the standard against randomly killing people. And we can act consistently with that standard by sanctioning him. In fact, it would be inconsistent for us to give him a pass.
Unless A was just asking to be walked through the calculation steps, then I agree B is not answering A’s question.
I’m not sure I’m following the argument here. I’m saying that all normativity is hypothetical. It sounds like you’re arguing there is a categorical ‘ought’ for believing mathematical truths because it would be very strange to say we only ‘ought’ to believe 2 + 2 = 4 in reference to some goal. So if there are some categorical ‘oughts,’ there might be others.
Is it something like that?
If so, then I would offer the goal of “in order to be logically consistent.” There are some who think moral oughts reduce to logical consistency, so we ought act in a certain way in order to be logically consistent. I don’t have a good counter-argument to that, other than asking to examine such a theory and wondering how being able to point out a logical consistency is going to rein in people with desires that run counter to it any better than relativism can.
You can stop right there. If no theory of morality based on logical consistency is offered, you don’t have to do any more.
I suppose you mean “if no theory of morality based on logical consistency is offered”.
Of course, one could make an attempt to research reason-based metaethics before discarding the whole idea.
Agreed and edited.
I observe that you didn’t offer a pointer to a theory of morality based on logical consistency.
I agree with Eby: you are a troll. I’m done here.
For one thing, I don’t think logical consistency is quite the right criterion for reason-based objective morality. Pointing out that certain ideas are old and well documented, is offering a pointer, and is not trolling.
I understand your point is that we can tell the killer that he has acted wrongly according to our standard (that one ought not randomly to kill people). But if people in general are bound only by their own standards, why should that matter to him? It seems to me I cannot provide him compelling grounds as to why he ought not to have done what he did, and that to punish him would be arbitrary. Sorry if I’m not getting it.
This states the thought very clearly -thanks.
I acknowledge the business about the nature of the compulsion behind mathematical judgement is pretty opaque. What I had in mind is illustrated by this dialogue. As it shows, the problem gets right back to the compulsion to be logically consistent. It’s possible this doesn’t really engage your thoughts, though.
If the view is correct, then you can at least convince rational people that it is not rational to kill people. Isn’t that an important result?
When a dispute is over fundamental values, I don’t think we can give the other side compelling grounds to act according to our own values. Consider Eliezer’s paperclip maximizer. How could we possibly convince such a being that it’s doing something irrational, besides pointing out that its current actions are suboptimal for its goal in the long run?
Thanks for the link to the Carroll story. I plan on taking some time to think it over.
It’s important to us, but — as far as I can tell — only because of our values. I don’t think it’s important ‘to the universe’ for someone to refrain from going on a killing spree.
Another way to put it is that the rationality of killing sprees is dependent on the agent’s values. I haven’t read much of this site, but I’m getting the impression that a major project is to accept this...and figure out which initial values to give AI. Simply ensuring the AI will be rational is not enough to protect our values.
that sounds like a good rational argument to me. Is the paperclip maximiser supposed to have a different rationality or just different values?
Like so much material on this site, that tacitly assumes values cannot be reasoned about.
If you don’t want murderers running around killing people, then it’s consistent with your values to set up a situation in which murderers can expect to be punished, and one way to do that is to actually punish murderers.
Yes, that’s arbitrary, in the same sense that every preference you have is arbitrary. If you are going to act upon your preferences without deceiving yourself, you have to feel comfortable with doing arbitrary things.
I think you missed the point quite badly there. The point is that there is no rationally compelling reason to act on any arbitrary value. You gave the example of punishing murderers, but if every value is equally arbitrary that is no more justifiable than punishing stamp collectors or the left-handed. Having accepted moral subjectivism, you are faced with a choice between acting irrationality or not acting. OTOH, you haven’t exactly given moral objectivism a run for its money.
I understand your point is that we can tell the killer that he has acted wrongly according to our standard (that one ought not randomly to kill people). But if people in general are bound only by their own standards, why should that matter to him? It seems to me I cannot provide him compelling grounds as to why he ought not to have done what he did, and that to punish him would be arbitrary.
This states the thought very clearly -thanks.
I acknowledge the business about the nature of the compulsion behind mathematical judgement is pretty opaque. What I had in mind is illustrated by this dialogue. As it shows, the problem gets right back to the compulsion to be logically consistent. It’s possible this doesn’t really engage your thoughts, though. Some people I know think it’s just foolish.
As is pointed out in the other thread from your post, plausibly our goal in the first instance is to show that it is rational not to kill people.
I don’t think that works. If you have multiple contradictory judgements being made by multiple standards, and you deem them to be objective, then you end up with multiple contradictoryobjective truths. But I don’t think you can have multiple contradictory objective truths.
You are tacitly assuming that the good guys are in the majority, However, sometimes the minority is in the right (as you and I would judge it), and need to persuade the majority to change their ways
It″ll work on people who already subscribe to rationaity, whereas relativism won’t.
What’s contradictory about the same object being judged differently by different standards?
Here’s a standard: return the width of the object in meters. Here’s another: return the number of wavelengths of blue light that make up the width of the object. And another: return the number of electrons in the object.
No Universally Compelling Arguments seems relevant here.
You realize that the linked post applies to arguments about mathematics or physics just as much as about morality.
Nothing. There’s nothing contradictory about multiple subjective truths or about multiple opinions, or about a single objective truth. But there is a contradiction in multiple objective truths about morality, as I said.
There isn’t any contradiction in multiple objective truths about different things; but the original hypothesis was multiple objective truths about the same thing, ie the morality of an action. If you are going to say that John-morality and Mary-morality are different things, that is effectively conceding that they are subjective.
The focus doesn’t have to be on John and Mary; it can be on the morality we’re referencing via John and Mary. By analogy, we could talk about John’s hometown and Mary’s hometown, without being subjectivists about the cities we are referencing.
That isn’t analogous, because towns aren;’t epistemic.
Hmm. Sounds like it would be helpful to taboo “objective” and “subjective”. Or perhaps this is my fault for not being entirely clear.
A standard can be put into the form of sentences in formal logic, such that any formal reasoner starting from the axioms of logic will agree about the “judgements” of the standard.
I should mention that this point that I use the word “morality” to indicate a particular standard—the morality-standard—that has the properties we normally associate with morality (“approving” of happiness, “disapproving” of murder, etc). This is the standard I would endorse (by, for example, acting to maximise “good” according to it) were I fully rational and reflectively consistent and non-akrasiac.
So the judgements of other standards are not moral judgements in the sense that they are not statements about the output of this standard. There would indeed be something inconsistent about asserting that other standards made statements about—ie. had the same output as—this standard.
Given that, and assuming your objections about “subjectivity” still exist, what do you mean by “subjective” such that the existence of other standards makes morality “subjective”, and this a problem?
It already seems that you must be resigned to your arguments failing to work on some minds: there is no god that will strike you down if you write a paperclip-maximising AIXI, for example.
Yep. Subjective statements about X can be phrased in objectivese. But that doesn’t make them objective statements about X.
By other standards do you mean other people’s moral standards, or non-moral (eg aesthetic standards)?
Of course. But I think moral objectivism is better as an explanation, because it explains moral praise and blame as something other than a mistake; and I think moral objectivism is also better in practice because having some successful persuasion going on is better than having none.
I don’t know what you mean, if anything, by “subjective” and “objective” here, and what they are for.
Okay… I think I’ll have to be more concrete. I’m going to exploit VNM-utility here, to make the conversation simpler. A standard is a utility function. That is, generally, a function that takes as input the state of the universe and produces as output a number. The only “moral” standard is the morality-standard I described previously. The rest of them are just standards, with no special names right now.
A mind, for example an alien, may be constructed such that it always executes the action that maximises the utility of some other standard. This utility function may be taken to be the “values” of the alien.
Moral praise and blame is not a mistake; whether certain actions result in an increase or decrease in the value of the moral utility function is a analytic fact. It is further an analytic fact that praise and blame, correctly applied, increases the output of the moral utility function, and that if we failed to do that, we would therefore fail to do the most moral thing.
By “subjective” I meant that it is indexed to an individual, and properly so. If Mary thinks vanilla is nice, vanilla is nice-for-Mary, and there is no further fact that can undermine the truth of that—whereas if Mary thinks the world is flat, there may be some sense in which it is flat-for-Mary, but that doens’t count for anything, because the shape of the world is not something about which Mary has the last word.
And there is one such standard in the universe, not one per agent?
If Mary thinks the world is flat, she is asserting that a predicate holds of the earth. It turns out it doesn’t, so she is wrong. In the case of thinking vanilla is nice, there is no sensible niceness predicate, so we assume she’s using shorthand for nice_mary, which does exist, so she is correct. She might, however, get confused and think that nice_mary being true meant nice_x holds for all x, and use nice to mean that. If so, she would be wrong.
Okay then. An agent who thinks the morality-standard says something other than it does, is wrong, since statements about the judgements of the morality-standard are tautologically true.
There is precisely one morality-standard.
Each (VNM-rational or potentially VNM-rational) agent contains a pointer to a standard—namely, the utility function the agent tries to maximise, or would try to maximise if they were rational. Most of these pointers within a light year of here will point to the morality-standard. A few of them will not. Outside of this volume there will be quite a lot of agents pointing to other standards.
Ok, instead of meter measurements, let’s look at cubit measurements. Different ancient cultures represented significantly different physical lengths by ‘cubits.’ So a measurement of 10 cubits to a Roman was a different physical distance than 10 cubits to a Babylonian.
A given object could thus be ‘over ten cubits’ and ‘under ten cubits’ at the same time, though in different senses. Likewise, a given action can be ‘right’ and ‘wrong’ at the same time, though in different senses.
The surface judgments contradict, but there need not be any propositional conflict.
Isn’t this done by appealing to the values of the majority?
Only if — independent of values — certain values are rational and others are not.
Are you sure that people mean different things by ‘right’ and ‘wrong’, or are they just using different criteria to judge whether something is right or wrong.
It’s done by changing the values of the majority..by showing the majority that they ought (in a rational sense of ought)) think differently. The point being that if correct reasoning eventually leads to uniform results, we call that objective.
Does it work or not? Have majorities not been persuaded that its wrong, if convenient, to oppress minorities?
What could ‘right’ and ‘wrong’ mean, beyond the criteria used to make the judgment?
Sure, if you’re talking about appealing to people to change their non-fundamental values to be more in line with their fundamental values. But I’ve still never heard how reason can have anything to say about fundamental values.
So far as I can tell, only by reasoning from their pre-existing values.
“Should be rewarded” and “should be punished”. If there was evidence of people saying that the good should be punished, that would indicate that some people are disagreeing about the meaning of good/right. Otherwise, disagreements are about criteria for assigning the term.
But not for all of them (since some of then get discarded) and not only from moral values (since people need to value reason to be reasoned with).
OK, but what I want to know is how you react to some person -whose belief system is internally consistent- who has just, say, committed a gratuitous murder. Are you committed to saying that there are no objective grounds to sanction him—there is no sense in which he ought not to have done what he did (assuming his belief system doesn’t inveigh against him offending yours)?
Touche.
Look, what I’m getting at is this. I assume we can agree that
“68 + 57 = 125” is true if and only if 68 + 57 = 125
This being the case, if A, seriously pondering the nature of the compulsion behind mathematical judgement, should ask, “Why ought I to believe that 68 + 57 = 125?”, and B answers, “Because it’s true”, then B is not really saying anything beyond, “Because it does”. B does not answer A’s question.
If the substantive answer is something along the lines that it is a mathematical fact, then I am interested to know how you conceive of mathematical facts, and whether there mightn’t be moral facts of generally the same ilk (or if not, why). But if you want somehow to reduce this to subjective goals, then it looks to me that mathematics falls by the wayside—you’ll surely allow this looks pretty dubious at least superficially.
Taking your thoughts out of order,
What I was getting at is that this looks like complete moral relativism -‘right for me’ is the only right there is (since you seem to be implying there is nothing interesting to be said about the process of negotiation which occurs when people’s values differ). I’m understanding that you’re willing to bite this bullet.
I take your point here. I may be conflating ethical and meta-ethical theory. I had in mind theories like Utilitarianism or Kantian ethical theory, which are general accounts of what it is for an action to be good, and do not aim merely to be accurate descriptions of moral discourse (would you agree?). If we’re talking about a defence of, say, non-cognitivism, though, maybe what you say is fair.
This is fair.
This is an interesting proposal, but I’m not sure what it implies. Is it possible for a rational person to strive to believe anything but the truth? Whether in math or anything else, doesn’t a rational person always try to believe what is correct? Or, to put the point another way, isn’t having truth as its goal part of the concept of belief? If so, I suggest this collapses to something like
*When considering 68 + 57, we are obligated to believe it equals 125 or some equivalent expression.
or, more plausibly,
*When considering 68 + 57, we ought to believe it equals 125 or some equivalent expression.
But if this is fair I’m back to wondering where the ought comes from.
That is an important point. People often run on examples as much as or more than they do on definitions,and if their intuitions about examples are strong, that can be used to fix their definitions (ie give them revised definitions that serve their intuitions better).
The rest of the post contained good material that needed saying.
When I have serious conversations with thoughtful religious people who have faith but no major theological training, I find it helpful to think of their statements about “God” as being statements about “all worldly optimization processes stronger than me that I don’t have time to understand in very much detail like evolution, entropy, economics, democratic politics, organizational dynamics, similar regularities in the structure of the world that science hasn’t started analyzing yet, plus many small activist groups throughout history, and a huge number of specific powerful agents silently influencing my life right now like various investors and celebrities, the local chief of police, the local school principal, my employer, my ancestors, and so on”.
I can imagine a relatively simple life heuristic, H, that might successfully navigate this vast and bewildering array of optimization pressures in their life and can ask “Does God want you to H”. Also, this translation scheme helps me to listen to evangelical radio and learn things from it :-)
I bring this up because it feels to me like you’re doing a lot of work to resuscitate ideas from moral philosophy that are significantly helped by “pluralistic reduction” to unpack the ideas into more specific and coherent claims, but you seem to be doing it in a lop-sided way by not unpacking the ideas of “the other side” in a similarly generous manner. Also, to a lesser extent, it seems to be leaving some of “our” ideas unpacked that could probably use some pluralistic reduction but might not look as shiny if unpacked this way.
I guess what I’m trying to say is that “acoustic messenger ferries” in the “ether” seem to me like perfectly adequate placeholder terms if I’m in a conversation with someone whose starting vocabulary uses them as atomic concepts, but if I’m tossing those terms out then “an ideally instrumentally rational and fully informed agent” seems roughly as questionable given how much difficulty people seem to have when using mind-shaped conceptually-atomic entities in their theories.
Do you think my impression of lopsided conceptual unpacking is accurate? If yes, I’m wondering if you could try to introspect on your writing process and try to articulate how you decided which things to unpack and which to leave fuzzy.
I’m not sure what you mean. I unpacked certain concepts as examples, and there are many more we could unpack. Could you say a bit more about what your concern is?
I don’t understand the terms “world of is” and “world of is not”. Does “talking about world of is not” mean “deducing from false assumptions”, or is there something more to it? Anyway, “talking about world of is” sounds like the worst kind of continental philosophy babble.
Else, the article is clear, comprehensible and well readable.
While “of is, of is not” didn’t hurt my understanding that much, the article would be better off without them.
I agree, and also note that the way luke dismisses the “is not” misses much of the point that is trying to be expressed by the phrase. If it is going to be discussed at all it deserves the same kind of parameterizing as ‘objective’ received.
It seems to be essentially a bit of wordplay, in that he uses it to mean two different things. Initially he is contrasting “is/is not” statements with “ought/ought not” statements. Later he talks about things that exist vs. things that don’t exist. It doesn’t seem to be very helpful though; in the earlier sense, there is no distinction between the “world of is” and the “world of is not”. So this seems like it was a bad idea.
I think there may be a good idea behind it though: view it as a cryptic appeal to Occam’s Razor. Various moralists (e.g., Railton, Craig) were shown to be speaking of real things and properties, or of imaginary ones, with their moral language. Why not then hypothesize that all are—albeit less transparently than these two—and do away with the need of a special metaphysics or semantics (or both) for “ought” questions as “opposed” to “is” questions.
Does this make it any clearer?
Yes, it does.
I just wanted to flag that a non-reductionist moral realist (like myself) is also “not here to debate definitions”. See my post on The Importance of Implications. This is compatible with thinking well of the Open Question Argument, if we think we have an adequate grasp of some fundamental normative concept (be it ‘good’, ‘reason’, or ‘ought’—I lean towards ‘reason’, myself, such that to speak of a person’s welfare is just to talk about what a sympathetic party has reason to desire for the person’s sake).
Note that if we’re right to consider some normative concepts to be conceptually primitive (not analytically reducible to non-normative concepts) then your practice of “tabooing” all normative vocabulary actually has the effect of depriving us of the conceptual tools necessary to even talk about the normative sphere. Consequent talk of people’s (incl. God’s) desires or dispositions is simply changing the subject, on this way of looking at things.
Out of interest: Will you be arguing anywhere in this sequence against non-reductionist moral realism? Or are you simply assuming its falsity from the start, and exploring the implications from there? (Even the latter, more modest project is of course worth pursing, but I personally would be more interested in the former.) Either way, it’d be good to be clear about this. (You could then skip the silly rhetoric about how what is not “is”, must be “is not”.)
I’m inclined not to write about moral non-naturalism because I’m writing this stuff for Less Wrong, where most people are physicalists.
What does it mean to you to say that something is a ‘fundamental normative concept’? As in… non-reducible to ‘is’ statements (in the Humean sense)?
I was thinking of “fundamental” concepts as those that are most basic, and not reducible to (or built up out of) other, more basic, concepts. I do think that normative concepts are conceptually isolated, i.e. not reducible to non-normative concepts, and that’s really the more relevant feature so far as the OQA is concerned. But by ‘fundamental normative concept’ I meant a normative concept that is not reducible to any other concepts at all. They are the most basic, or bedrock, of our normative concepts.
Given the extremely poor access human beings have to the structure of their own concepts, it’s dubious that the methods of analytic philosophy can trace those structures. Moreover, concepts typically “cluster together similar things for purposes of inference” ( Yudkowsky ) and thus we can re-structure them in light of new discoveries. Concepts that are connected now might be improved by disconnecting them, or vice versa. It is not at all clear that normative concepts are not included in this (Neurath-style) boat.
Physicalists could (like Mackie) accept the non-naturalist’s account of what it would take for something to be genuinely normative, and then simply deny that there are any such properties in reality. I’m much more sympathetic to this hard-headed “error theory” than to the more weaselly forms of naturalism.
I think many of our normative concepts fail to refer, but that a class of normative concepts often called hypothetical imperatives do refer, thanks to a rather straightforward reduction as given above. Are hypothetical imperatives not ‘genuinely normative’ in your sense of the phrase? Do you use the term ‘normative’ when talking about things other than hypothetical imperatives, and do you think those other things successfully refer?
As I argue elsewhere:
“Hypothetical imperatives thus reveal patterns of normative inheritance. But their highlighted ‘means’ can’t inherit normative status unless the ‘end’ in question had prior normative worth. A view on which there are only hypothetical imperatives is effectively a form of normative nihilism—no more productive than an irrigation system without any water to flow through it.”
(Earlier in the post explains why hypothetical imperatives aren’t reducible to mere empirical statements of a means-ends relationship.)
I tentatively favour non-naturalist realism over non-naturalist error theory, but my purpose in my previous comment was just to flag the latter option as one that physicalists should take (very) seriously.
Error theory
You know this, but for the benefit of others: Roughly, error theory consists of two steps. As Finlay puts it:
Given my view of conceptual analysis, it shouldn’t be surprising that I’m not confident of some error theorists’ assertion of step 1. Is a presupposition of moral absolutism ‘essential’ to a judgment’s status as a ‘moral’ judgment? Is a presuppositional of motivational internalism ‘essential’ to a judgment’s status as a ‘moral’ judgment? I don’t know. Moral discourse (unlike carbon discourse) is so confused that I’m not too interested to assert one fine boundary line around moral terms over another.
So if someone thinks a presupposition of supernaturalism is ‘essential’ to a judgment’s status as a ‘moral’ judgment, then I will claim that supernaturalism is false. But this doesn’t make me an error theorist because I don’t necessarily agree that a presupposition of supernaturalism is ‘essential’ to a judgment’s status as a ‘moral’ judgment. I reject step 1 of error theory in this case.
Likewise, if someone thinks a presupposition of moral absolutism or motivational internalism is essential to a judgment’s status as a ‘moral’ judgment, I’ll be happy to deny both moral absolutism and motivational internalism, but I wouldn’t call myself an error theorist because I reject the claim that moral judgments (by definition, by conceptual analysis) necessarily presuppose moral absolutism or motivational internalism.
But hey, if you convince me that the presumption of motivational internalism in moral discourse is so widespread that talking about ‘morality’ without it would be like using the term ‘phlogiston’ to talk about oxygen, then I’ll be happy to call myself an error theorist, though none of my anticipations will have changed.
Hypothetical imperatives
I’ll reply to a passage from your post on hypothetical imperatives. My reply won’t make sense to those who haven’t read it:
I think this is because ‘should’ is being used in different senses. The real modus ponens is:
If you want to torture children, you should_ToTortureChildren volunteer as a babysitter.
You want to torture children.
Therefore, you should_ToTortureChildren volunteer as a babysitter.
Or at least, that is plausibly what some people mean when they assert what looks like a hypothetical imperative. Doubtless, others will appear to be meaning something else if pressed by interrogation.
Now, to respond to Sidgwick:
I could capture this ‘unreasonableness’ by simply clarifying that from the standpoint of Bayesian rationality, it would be somewhat irrational to expect good health despite not rising early (or so the doctor claims).
But again, I’m not too keen to play the definitions game. If you state hypothetical imperatives with more intuitions about the meaning of hypothetical imperatives than I do, then you are free to explain what you mean by hypothetical imperatives and then show how they fit into the physical world. If you can’t show how they fit into the world, then you’re talking about something that doesn’t exist, or else we’ll have to replay the physicalism vs. non-physicalism debate, which is another topic.
Right?
Thanks for this reply. I share your sense that the word ‘moral’ is unhelpfully ambiguous, which is why I prefer to focus on the more general concept of the normative. I’m certainly not going to stipulate that motivational internalism is true of the normative, though it does seem plausible that there’s something irrational about someone who acknowledges that they really ought (all things considered) to phi and yet fails to do so. (I don’t doubt that it’s possible for someone to form the judgment without any corresponding motivation though, as it’s always possible for people to be irrational!)
I trust that we all have a decent pre-theoretic grasp of normativity (or “ought-ness”). The question then is whether this phenomenon that we have in mind (i) is reducible to some physical property, and (ii) actually exists.
Error theory (answering ‘no’ and ‘no’ to the two questions above) seems the most natural position for the physicalist. And it sounds like you may be happy to agree that you’re really an error theorist about normativity (as I mean it). But then I’m puzzled by what you take yourself to be doing in this series. Why even use moral/normative vocabulary at all, rather than just talking about the underlying natural properties that you really have in mind?
P.S. What work is the antecedent doing in your conditional?
Why do you even need the modus ponens? Assuming that “should_ToTortureChildren” just means “what follows is an effective means to torturing children”, then isn’t the consequent just plain true regardless of what you want? (Perhaps only someone with the relevant desire will be interested in this means-ends fact, but that’s true of many unconditional facts.)
Right.
Unfortunately, I don’t think I’m clear about what you mean by normativity. The only source of normativity I think exists is the hypothetical imperative, which can be reduced to physics by straightforward methods such as the one I used in the original post. I’m not an error theorist about that kind of normativity.
This is a good question. Truly, I want to get away from moral vocabulary, and be careful around normative vocabulary. But people already think about these topics in moral and normative vocabulary, which is why I’m trying to solve or dissolve (in this post and its predecessor) some of the usual ‘problems’ in this space of questions.
After that’s done, I don’t think it will be most helpful to use moral language. This is evident in the fact that in 15 episodes of my ‘morality podcast’ I’ve used almost no moral language at all.
Not much, really. I wasn’t using the modus ponens to present an argument, but to unpack one interpretation of (some) ‘should’ discourse. Normative language, like many other kinds of language, is (when used correctly) merely a shortcut for saying something else. I can imagine a language that has no normative language at all. In that language we couldn’t say things like “If you want to torture children, you should volunteer as a babysitter” but we could say things like “If you volunteer as a babysitter you will have more opportunities to torture children.” The way I’m parsing ‘should’ in the first sentence, nothing is lost by this translation.
Of course, people use ‘should’ in a variety of ways, some of which translate into claims about things reducible to physics, others of which translate into claims about things non-reducible to physics, while still others don’t seem to translate into cognitive statements at all.
Thanks, this is helpful. I’m interested in your use of the phrase “source of normativity” in:
This makes it sound like there’s a new thing, normativity, that arises from some other thing (e.g. desires, or means/ends relationships). That’s a very realist way of talking.
I take it that what you really want to say something more like, “The only kind of ‘normativity’-talk that’s naturalistically reducible and hence possibly true is hypothetical imperatives—when these are understood to mean nothing more than that a certain means-end relation holds.” Is that right?
I’d then understand you as an error theorist, since “being a means-end relationship”, like “being red”, is not even in the same ballpark as what I mean by “being normative”. (It might sometimes have normative importance, but as we learn from Parfit, that’s a very different thing.)
My thought process on sources of normativity looks something like this:
People claim all sorts of justifications for ‘ought’ statements (aka normative statements). Some justify ought statements with respect to natural law or divine commands or non-natural normative properties or categorical imperatives. But those things don’t exist. The only justification of normative language that fits in my model of the universe is when people use ‘ought’ language as some kind of hypothetical imperative, which can be translated into a claim about things reducible to physics. There are many varieties of this. Many uses of ‘ought’ terms can be translated into claims about things reducible to physics. If somebody uses ‘ought’ terms to make claims about things not reducible to physics, then I am suspicious of the warrant for those claims. When interrogating about such warrants, I usually find that the only evidences on offer are pieces of folk wisdom, intuitions, and conventional linguistic practice.
You still seem to be conflating justification-giving properties with the property of being justified. Non-naturalists emphatically do not appeal to non-natural properties to justify our ought-claims. When explaining why you ought to give to charity, I’ll point to various natural features—that you can save a life for $500 by donating to VillageReach, etc. It’s merely the fact that these natural features are justifying, or normatively important, which is non-natural.
Sure. So what is it that makes (a) [the fact that you can save a life by donating $500 to VillageReach] normatively justifying, whereas (b) [the fact that you can save a mosquito by donating $2000 to SaveTheMosquitos] is not normatively justifying?
On my naturalist view, the fact that makes (a) but not (b) normatively justifying would be some fact about how the goal we’re discussing at the moment is saving human lives, not saving mosquito lives. That’s a natural fact. So are the facts about how the English language works and how two English speakers are using their terms.
It’s not entirely clear what you’re asking. Two possibilities, corresponding to my above distinction, are:
(1) What (perhaps more general) normatively significant feature is possessed by [saving lives for $500 each] that isn’t possessed by [saving mosquitoes for $2000 each]? This would just be to ask for one’s fully general normative theory: a utilitarian might point to the greater happiness that would result from the former option. Eventually we’ll reach bedrock (“It’s just a brute fact that happiness is good!”), at which point the only remaining question is....
(2) In what does the normative signifiance of [happiness] consist? That is, what is the nature of this justificatory status? What are we attributing to happiness when we claim that it is normatively justifying? This is where the non-naturalist insists that attributing normativity to a feature is not merely to attribute some natural quality to it (e.g. of “being the salient goal under discussion”—that’s not such a philosophically interesting property for something to have. E.g., I could know that a feature has this property without this having any rational significance to me at all).
(Note that it’s a yet further question whether our attributions of normativity are actually correct, i.e. whether worldly things have the normative properties that we attribute to them.)
I gather it’s this second question you had in mind, but again it’s crucial to carefully distinguish them since non-naturalist answers to the first question are obviously crazy.
Yup. I’m asking question (2). Thanks again for your clarifying remarks.
What if you actually should be discussing saving of mosquito lives, but don’t, because humans are dumb?
I think this is a change of subject, but… what do you mean by ‘actually should’?
No idea.
I take you to mean “what would maximize Luke’s utility function” (knowing that ‘utility function’ is probably just a metaphor when talking about humans) when you say “you actually should...” Of course, my ‘utility function’ is unknown to both of us.
In that case, it would remain true in our hypothetical scenario that I should-HumanLivesAreGood donate to VillageReach (assuming they’re a good charity for saving human lives), while I should-UtilityFunctionLuke donate to SaveTheMosquitos.
(Sorry about formatting; LW comments don’t know how to use underscores, apparently.)
But the question then is what goal you should have. It is easy to naturalise norms inasmuch as they are hypothetical and indexed to whatever you happen to be doing. (if you want to play chess, you should move the bishop diagonally) The issue is how to naturalise categorical ends,the goals you should have and the rules you should be following irrespective of what you are doing.
Such facts aren’t supernatural. OTOH, they fall on the analytical/apriori side of the fence, rather than the empirical side, and that is an iimportant distinction.
What reasons are there for doubting the existence of categorical imperatives that do not equally count against the existence of hypothetical imperatives? I can understand rejecting both, I can understand accepting both, but I can’t understand treating them differently.
Depends what you mean by ‘categorical imperative’, and what normative force you think it carries. The categorical imperatives I am used to hearing about can’t be reduced into the physical world in the manner by which I reduced a hypothetical imperative into the physical world in the original post above.
I take it you want to reduce a hypothetical imperative like “If you want to stretch, then you ought to stand” into a physically-kosher causal claim like “standing is causally necessary for satisfying your desire to stretch”. Now, I’m skeptical about this reduction, simply because I don’t see how a mere causal claim could provide any normative direction whatsoever. But in any case, it seems that you could equally well reduce a categorical imperative like “Regardless of what you personally want, if you are the only one around a drowning child, then you ought to help save it from drowning” into a physically-kosher causal claim like “your help is causally necessary for the survival of the drowning child”. Both causal claims are equally physicalistic, and both seem equally relevant to their respective imperative, so both physicalistic reductions seem equally promising.
Of course, the question “why should I assume that the drowning’s child survival actually matters?” seems reasonable enough. But so does the question “why should I assume that satisfying my desire to stretch actually matters?” If the first question jeopardizes the reduction of the categorical imperative, then the second question would also jeopardize the reduction of the hypothetical imperative.
I’m not used to the term ‘categorical imperative’ being used in this way. Normally, a categorical imperative is one that holds not just regardless of what you want, but regardless of any stated ends at all. Your reduction of this categorical imperative assumes an end of ‘survival of the drowning child.’ To me, that’s what we call a hypothetical imperative, and that’s why it can be reduced in that way, by translating normativity into statements about means and ends.
I think you must be mistaken about categorical imperatives and ends.
In the Groundwork, Kant deems rational nature (‘humanity’ in us) to be an end in itself which can ground a categorical imperative, in the 2nd Critique, he deems the highest good (happiness in proportion to virtue) to be a necessary end of practical reason, and in the Metaphysics of Morals, he deems one’s own perfection and the happiness of others to be ends that are also duties. For Kant, the categorical imperative of morality is directed at ends, but it is not a mere hypothetical imperative grounded in subjective ends.
And leaving Kant aside, plenty of moral systems aspiring to categorical force have ends at their center. Classic utilitarianism picks out the greatest overall balance of pleasure over pain as the one and only end of morality. Eudaimonist virtue ethics has as its end a well-lived flourishing life for the agent. Thomist ethics says an intellectual vision of God in the afterlife is the ultimate end of humans. These systems do not traffic in merely hypothetical imperatives: they present these ends as objectively worth pursuing, regardless of the agent’s personal preferences.
But if your main point is simply that my ‘reduction’ assumes that the end ought to be pursued, and that this assumption is a form of cheating, then I agree. I’ve left the normativity unexplained and unreduced. But then in exactly the same way, your reduction of hypothetical imperatives assumes that effective means to one’s ends ought to be taken, and this assumption is also a form of cheating. You have also left the normativity unexplained and unreduced.
So I don’t see how hypothetical imperatives are any more fit for naturalistic reduction than categorical imperatives.
I’m sure I speak differently about categorical imperatives than Kant does. I haven’t read much of Kant, and I don’t regret that. In your language, what you call an “end in itself” is what I mean to pick out when I talk about an imperative that “holds… regardless of any stated [arbitrary, desired, subjective] ends at all.” I don’t really know what it means for something to be an “end in itself”. Kant’s idea seemed to be that we ought to do X regardless of what anyone wants. Your way (and perhaps Kant’s way) of talking about this is to say that we ought to do X regardless of what anyone wants because X leads to a particular “end in itself”, whatever that means.
My reduction of hypothetical imperative doesn’t assume this. It only translates ‘ought’ into a prediction about what would realize the specified end. If there’s something mysterious left over, I’m curious what you think it is and whether it is real or merely a figment of folk wisdom and linguistic practice.
So, on your use of ‘end’, an ‘end’ cannot be objective and unconditional? I think that’s a highly uncommon use of the term.
But if you go this way, it seems like it’s less of a reduction of ‘ought’ and more of a misinterpretation, like reducing ‘Santa Claus’-talk into talk about Christmas cheer, or ‘God’-talk into talk of love.
After all, one important constraint on any interpretation of any ‘ought to X’ is that it should be positive towards X as opposed to negative or neutral, in favor of some action or attitude as opposed to against it or indifferent. But a mere predictive causal claim doesn’t have any valence at all: it’s just a neutral claim about what will probably lead to what, without anything positive or negative. So any attempt to reduce oughts to predictive causal claims seems doomed to failure.
EDIT: For the record, I’m an expressivist about normativity, and I think any attempt to understand it in terms of some actual or hypothetical ontology that could serve as the truth-conditions for a descriptive belief is a mistake. The mystery, I would say, lies in a descriptive interpretation of normativity, not in normativity itself.
No, I just should have been clearer that when I said “stated end” I meant “subjective, desired end”. As commonly used, ‘end’ includes unconditional ends, I just haven’t ever been presented with an argument that persuaded me to think that such ends exist.
Sure, you can stick that feature into your meaning of ‘ought’ if you want. But I’m not going too deep into the conceptual analysis game with you. If you want to include positive valence in the meaning of a hypothetical ought, then we could translate like this:
“If you want to stretch, you ought to stand” → “standing will make it more likely that you fulfill your desire to stretch, and I have positive affect toward you standing so as to fulfill your desire to stretch”
As I said in my post, expressivism is fine—and certainly true of how many speakers use normative language. I happen to be particularly interested in those who use normative language cognitively, and whether their stated moral judgments are true or false, and under which conditions—which is why I’m investigating translations/reductions of normative language into natural statements that have truth conditions.
What do you think human wellbeing is for,then?
The set of non-ethical categorical imperatives is non-empty. The set of non-ethical hypothetical imperatives is non-empty. Hypothetical imperatives include instrumental rules, you have to use X to achieve Y, game-laying rules, etc.
How exactly does this answer the question?
I agree. Epistemic imperatives are categorical, but non-empty.
Right, those are examples where non-ethical hypothetical imperatives often show up.
So how does this add up to a reason to think there is a case against categorical imperatives that doesn’t equally well count against hypothetical imperatives?
My thought process on sources of normativity looks something like this:
There’s a lot of kinds of normative/”ought” statements. Some relate to games, some to rationality, and so on. Hypothetical “ought” statements do not require any special metaphysical apparatus to explain them, they just require rules and payoffs. Categorical imperatives are another story.
One man’s conventional linguistic practice is another’s analytical truth.
Rules and payoffs explain “ought” statements only if you assume that the rules are worth following and the payoffs worth pursuing. But if hypothetical imperatives can help themselves to such assumptions (assuming e.g. that one’s own desires ought to be satisfied), then categorical imperatives can help themselves to such assumptions (assuming e.g. that everyone’s desires ought to be satisfied, or that everyone’s happiness ought to be maximized, or that everyone ought to develop certain character traits).
I don’t think so. You ought to use a hammer to drive in nails even if you don’t want to dive in nails. Anyone who is playing chess should move the bishop diagonally.That doesn’t mean you are playing chess.
Of course those are hypothetical, and non-ethical. It might wll be the case that the only categorical imperatives are moral categorical imperatives; that. ethics is the only area where you should do things or refrain form things unconditionally.
Again, you’re assuming that the rule ‘if you’re driving in nails, use a hammer’ is worth following, and that the rule ‘if you’re playing chess, move bishops diagonally’ is worth following. A nihilist would reject both of those rules as having any normative authority, and say that just because a game has rules it doesn’t mean that game-players ought to follow those rules, at most it means that lots and lots of rule-violations make the game go away.
I don’t think hypothetical imperatives can be reduced. The if-ought of a hypothetical imperative is a full-blooded normative claim. But you can’t reduce that to a simple if-then about cause and effect.
To see why, consider a nihilist about oughts. She recognizes the causal connections between calorie consumption/burning and weight loss. But she doesn’t accept any claim about what people ought to do, even hypothetical imperatives about people who desire weight loss. This seems perfectly coherent: she accepts causal claims, but not normative claims, and there’s no contradiction or incoherence there. But this means the causal claims she accepts are not conceptually equivalent to the normative claims she rejects.
For another way to see why, consider the causal claim “less calorie consumption and more calorie burning leads to weight loss”. This causal claim points in no normative direction. It doesn’t recommend anything, or register any approval, or send any positive or negative messages. Of course, we can take it in one direction or another, but only by combining it with separate normative claims:
Less calorie consumption and more calorie burning leads to weight loss.
People ought to take causally efficacious steps to satisfy their desires.
Therefore, if you desire to lose weight, you ought to consume less calories and burn more calories.
Premise 2 is what provides the normativity. It points us in the direction of satisfying desires. But we could easily take things in the opposite direction.
Less calorie consumption and more calorie burning leads to weight loss.
People ought to take causally efficacious steps to frustrate their desires.
Therefore, if you desire to lose weight, you ought to consume more calories and burn less calories.
Again, premise 2 is what provides the normativity. But it points us in the opposite direction, viz. the direction of frustrating desires.
So it’s pretty clear that premise 1 has no normativity in it. It can’t be reduced to either of the two 3′s. For we cannot arrive at a 3 without a 2.
I think stating premise 2 is a little odd. It is a bit deja “Tortoise and Achilles” all over again. If there’s a norm hiding around here, it’s an “ought” portrayed by the desire.
Second, conceptual analysis (or conceptual equivalence) is not necessary for reduction. Look at reduction in the sciences for examples.
Well, I’ll acknowledge that you could change premise 2 into an inference rule. But notice that you could change either premise 2—the pro-desire-satisfaction one and the pro-desire-frustration one—into an inference rule. Indeed, you could change any normative claim into an inference rule: you could change “people who want to have gay sex ought to go see a trained Baptist minister to get cured” into an inference rule, and then validly go from “I want to have gay sex” to “I ought to go see a trained Baptist minister to get cured”. So from the fact that premise 2 could be changed into an inference rule, I don’t think anything follows that might jeopardize its status as a full-blooded normative claim.
On the second point, I thought lukeprog was discussing direct conceptual reduction. But if he wants to provide hypothetical imperatives with a synthetic reduction, he’ll need a theory of reference capable of explaining why the normative claim turns out to make reference to (and have its truth-conditions provided by) simple causal facts. And on this score, I think hypothetical imperatives and categorical moral imperatives are on an equal footing: since reductionist moral realists have a hard time with synthetic reductions, I would expect reductionist ‘instrumental realists’ to have a hard time as well.
For what it’s worth, I think what’s really being inferred by the advice-giver is:
1 Granting your (advisee’s) starting point, you ought to lose weight.
2 Less calorie consumption and more calorie burning leads to weight loss.
3 Therefore you ought to consume less and burn more calories.
The advisee’s desire portrays the starting-point as a truth.
Perhaps so, but then the normativity stems from premise 1, leaving premise 2 as non-normative as ever. But the question is whether premise 2 could be a plausible reduction basis for normative claims.
Tangentially:
How so? (Note that a proposition may be in some sense about minds without its truth value being mind-dependent. E.g. “Any experience of red is an experience of colour” is true regardless of what minds exist. I would think the same is true of, e.g., “All else equal, pain is bad for the experiencer.”)
I’m borrowing the concept ‘well-being of conscious creatures’ from Sam Harris, who seems to think of it in terms of mind-dependent facts, perhaps involving (e.g.) brain states we might call ‘pain’ or ‘pleasure’.
That doesn’t really answer my question. Let me try again. There are two things you might mean by “mind dependent”.
(1) You might just mean “makes some reference to the mind”. So, for example, the necessary truth that “Any experience of red is an experience of colour” would also count as “mind-dependent” in this sense. (This seems a very misleading usage though.)
(2) More naturally, “mind dependent” might be taken to mean that the truth of the claim depends upon certain states of mind actually existing. But “pain is bad for people” (like my example above) does not seem to be mind-dependent in this sense.
Which did you have in mind?
By saying that “facts about the well-being of conscious creatures are mind-dependent facts,” I just mean that statements about the well-being of conscious creatures are made true or false by facts abound mind states. A statement about my well-being is mind-dependent in the sense that a statement about my well-being (as I am using the term) is a statement about my brain states. A statement about the distance between my chair and my desk is not a statement about brain states, and would be true or false whether or not our Hubble volume still contained any minds.
I believe the “facts” in question were synthetic ones (“all else being equal, being set on fire is bad for the person set on fire,”) not analytic ones (“all else equal, pain is bad for the experiencer.”)
It’s not analytic that pain is bad. Imagine some crazy soul who thinks that pain is intrinsically good for you. This person is deeply confused, but their error is not linguistic (as if they asserted “bachelors are female”). They could be perfectly competent speakers of the english language, and even logically omniscient. The problem is that such a person is morally incompetent. They have bizarrely mistaken ideas about what things are good (desirable) for people, and this is a substantive (synthetic), not merely analytic, matter.
Perhaps the thought is that contingent (rather than necessary) facts about wellbeing are mind-dependent. That’s still not totally obvious to me, but it does at least seem less clearly false than the original (unrestricted) claim.
The issue is more whether anyone could think pain is good for them themselves. One could imagine a situation where pain receptors connect up to pleasure centers, but then it becomes a moot point as to whether that is actually pain.
Yes, I was imagining someone who thought that unmitigated pain and suffering was good for everyone, themselves included. Such a person is nuts, but hardly inconceivable.
In the not distant past, some surgeons opposed pain-killing medication for post-operative pain, believing that the pain was essential to the healing process.
There’s also the reports by patients who have had morphone for pain relief, that the pain is still there, but it takes the hurting out of it.
Just to clarify: By ‘pain’ I mean the hurtful aspect of the sensation, not the base sensation that could remain in the absence of its hurting.
In your first paragraph you describe people who take pain to be instrumentally useful in some circumstances, to bring about some other end (e.g. healing) which is itself good. I take no stand on that empirical issue. I’m talking about the crazy normative view that pain is itself (i.e. non-instrumentally) good.
I’m with RichardChappell. Deeming pain intrinsically good is pretty easy, though highly unusual. And deeming pain intrinsically value-neutral is even easier, and not all that unusual. Thus there’s nothing incoherent about denying that pain is intrinsically bad.
For example, plenty of people think animal pain doesn’t matter, not intrinsically anyway. Perhaps Kant is the most famous example. He thinks cruelty to animals is morally wrong only because it is likely to make us cruel to humans. But the animal pain itself is, Kant thinks, irrelevant and perfectly devoid of any value or disvalue.
Certain Stoics would even say that human pain doesn’t matter, not intrinsically anyway. It matters only inasmuch as it causally relates to one’s own virtue, and it has no intrinsic relevance to what is good.
If someone went even further, reversing common sense and insisting that pain were intrinsically good, that would be unusual. But it wouldn’t be incoherent. Not even close. To invent an example, suppose an extremely credulous religious person were told by their leader that pain is intrinsically good. This true-believer would then be convinced that pain is intrinsically good, and they would try to bring about pain in themselves and in others (so long as they didn’t violate any other moral rules endorsed by the leader).
Assume you have a complete knowledge of all details of the way human brain works, and a detailed trace of the sequence of neurological events that leads people to ask moral questions. Then what?
My only guess is that you look this trace over using your current moral judgment, and decide that you expect that changing certain things in the algorithm will make the judgments of this brain better. But this is not a FAI-grade tool for defining morality (unless we have to go the uploads-driven way, in which case you just gradually and manually improve humans for a very long time).
Yes, a completed cognitive neuroscience would certainly not be sufficient for defining the motivational system of an FAI.
I think you have hit the nail on the head. There may surely be lots of scientifically interesting and useful reasons for investigating the fine details of the brain processes which eventuate in behaviours -or the uttering of words- which we interpret as moral, but it is far from obvious that this kind of knowledge will advance our understanding of morality.
More generally, there is plausibly a tension (at least on the surface) between two dominant themes on this site:
1) Naturalism: All knowledge—including of what’s rational (moral)- is scientific. To learn what’s rational (moral) our only option is to study our native cognitive endowments.
2) Our/evolution’s imperfection: You can’t trust your untutored native cognitive endowment to make rational (or moral) judgements. Unless we make an effort not to, we make irrational judgements.
Strictly speaking, we can exhibit any definition of “good”, even one that doesn’t make any of the errors you pointed out, and still ask “Is it good?”. The criteria for exhibiting a particular definition are ultimately non-rigorous, even if the selected definition is, so we can always examine them further.
Moore’s argument might fail in the unintended use case of post-FAI morality not because at some point there might be no more potential for asking the question, but because, as with “Does 2+2 equal 4?”, there is a point at which we are certain enough to turn to other projects, even if in principle some uncertainty and lack of clarity in the intended meaning remains. It’s not at all clear this will ever happen to morality.
Moore was correct that no alternate concrete meaning is identical to ‘good,’ his mistake was thinking that ‘good’ had its own concrete meaning. As Paul Ziff put it, good means ‘answers to an interest’ where the interest is semantically variable.
In math terms, the open question argument would be like asking the value of f(z) and when someone answers f(3), pointing out that f(z) is not the same thing as f(3).
I think the ‘huge and complicated’ X that Luke mentions is supposed to be the set all of inputs to f(z) that a given person is disposed to use. Or maybe the aggregate of all such sets for people.
In The Is_Ought Gap, Luke writes
Ironically, this is where I quickly lost interest in this article, because glib word-play isn’t my subject of interest.
I wasn’t trying to be glib. All I’m saying is that if somebody uses ‘ought’ terms to refer to things that can’t be reduced to physics, then they’re either talking about things that don’t exist or they aren’t physicalists—and physicalism vs. non-physicalism is beyond the scope of this particular article.
Can “the best way to play chess” be reduced to physics? Can it usefully be rduced to physics.
Your emphasis on can’t. But do you really mean “can’t be reduced”? Or are you rather excluding anything that “hasn’t been reduced”? In the original posting you seem to be demanding that anyone using “ought language” be prepared to perform the reduction on the spot:
I’m sorry, Luke. My commitment to reductionism is more of a vague hope for future physicalist explanation. I did not agree to use only that language which I can simultaneously show to be universally tabooable.
Perplexed,
I feel you’re being uncharitable. Do you really think we’re having a disagreement about how reductionism works? I’m not demanding that somebody have a full down-to-atoms reduction in mind whenever they use any term whatever. My target in that paragraph is people who themselves don’t think their use of ‘ought’ will ever reduce into physics—for example Richard Chappell.
I believe that you do feel that. But if you think that excluding people like Chappell from the discussion is a fruitful way to proceed, then I am curious why you believe that what you are discussing is properly termed “Meta-ethics”.
ETA: Ok, I’ll try for a bit more charity. Why does it matter whether a reduction of “ought” to physics can or cannot be accomplished? Why not simply present your ideas and then point out that they have the virtue of making ‘ought’ reducible?
It seems to me that you are trying too hard to be a good follower of Eliezer. Take my word for it; that isn’t necessary to be respected in this forum. It is possible (though admittedly sometimes expensive) to communicate without first achieving some kind of metaphysical solid ground.
I didn’t spend more time on reductionism because that is covered in the reductionism sequence. This post is already too long; I don’t have the space to re-hash all the arguments for reductionism. Chappell already knows what’s in the reductionism sequence, and why he disagrees with it, and that is a different discussion.
Luke’s position seems more reasonable here.
I have to say that the “following” around here is kinda irritating to me too.
There’s a scene from The Life of Brian that comes to mind—the “you’ve got to think for yourselves” scene.
Can you elaborate?
“World of is” vs “World of is not” is a false dichotomy.
“I lose interest” is possibly the worst and most uncharitable of all forms of philosophical rhetoric. (Hence my ‘turnabout’.)
Luke gave a false impression of what non-naturalistic ethics comprises by providing only a single example of such a position—a position which was absurd because it presumed the existence of a deity.
Even Hume allowed that works of mathematics need not be “committed to the flames”. Mathematics, to my mind, does not deal with the “world of is”, neither does it deal with the “world of is not”. Yet if someone were to provide a non-reductionist, but axiomatic, ‘definition’ of ‘good’ using the methods and practices of mathematical logic, I certainly would not dismiss it as “uninteresting”.
But mathematics does deal with the “world of is” either potentially or as “rules of thought” (all thoughts are in minds). God, on the other hand, is different.
In my mind, the practice of mathematics is the practice of distinguishing between the “world of is” and the “world of is not;” if something is mathematically provable, it is; if something is mathematically disprovable it is not.
Existence and non-existence pretty much exausts all possibilities here, no?
Consider this dialog:
Student: “Wise master, what ought I do?”
Wise master: “You ought to help the poor by giving 50% of your income to efficient charity and supporting the European-style welfare state.”
Student: “Alright.”
*student runs off and gives 50% of his or her income to efficient charity and supports the European-style welfare state
This dialog rings true as a fact about ought statements—once we become convinced of them, they do and should constrain our behavior.
But my dialogs and your dialogs contradict each other! Because if “ought” determines our behavior, and we can define what “ought” means, then we can define proper behavior into existence—a construction as absurd as Descartes defining God into existence or Plato defining man as both a hairless featherless biped and a mortal.
We must give up one, and I say give up yours. “ought” is one of those words that we are not free to define—it has a single meaning. Look to its consequences, not its causes.
I’m not sure I understand. Are you saying that we are not free to stipulate definitions for the word-tools we use (when it comes to morality), because you have a conceptual intuition in favor of motivational internalism for the use of ‘ought’ terms?
Wikipedia defines motivational internalism as the belief that:
I want to view this as a morally necessary connection. One should do what one ought to do, and this serves as the definition of “ought”.
You will note that I am using circular definitions. That is because I can’t define “should” except in terms of things that have a hidden “should” in there. But I am trying to access the part of you that understands what I am saying.
The useful analogue is this:
modus ponens: “If you know ‘A’, and you know ’If A, then B”, then you know B”
It’s a circular definition getting at something which you can’t put into words. I would be wrong to define “If-then” as something else, like maybe “If A, then B” means “75% of elephants with A written on them believe B” because it’s already defined.
Does that make any sense?
Unfortunately, I still don’t follow you. Or at least, the only interpretations I’ve come up with look so obviously false that I resist attributing them to you. Maybe I can grok your disagreement from another angle. Let me try to pinpoint where we disagree. I hope you’ll have some time to approach mutual understanding on this issue. When Will Sawin disagrees with me, I pay attention.
Do you agree that there are many words X such that X is used by different humans to mean slightly different things?
Do you agree that there are many words X such that different humans have different intuitions about the exact extension of X, especially in bizarre sci-fi hypothetical scenarios?
Do you agree that many humans use imperative terms like ‘ought’ and ‘should’ to communicate particular meanings, with these meanings often being stipulated within the context of a certain community?
I’ll stop there for now.
Thanks. I’m thinking of doing a post on the discussion section where I can explain where my intuitions come from in more detail.
For your questions:
Yes.
Yes.
I don’t really know what the third question means. It seems like the primary use of “ought” and “should” is as part of an attempt to convince people to do what you say they should do. I would say that is the meaning being communicated. There are various ways this could be within the context of a community. Are you saying that you’re only trying to convince members of that community?
Note: I’m planning to come back to this discussion in a few days. Recently my time has been swamped running SI’s summer minicamp.
I may also write something which expresses my ideas in a new, more concise and clear form.
I think that would be the most efficient thing to do. For now, I’ll wait on that.
If you haven’t noticed, I just made that post.
Any response to this?
Excellent. I’m busy the next few days, but I’ll respond when I can, on that thread.
I think you meant to leave out either the “except” or the “don’t”?
Correct.
Your dialogue looks similar to the one about losing weight above. I can define proper behavior given my terminal values. If I want to lose weight, I should eat less. Upon learning this fact, I start eating less. My values and some facts about the world are sufficient to determine my proper behavior. “Defining my behavior into existence” seems no more absurd to me than defining the rational action using a decision theory.
I’m not sure I’ve explained myself very clearly here. Please advise on what, if anything, that I’m saying is hard to understand.
If it is the case that you should do what you want, yes.
If you want to punch babies, then you should not punch babies. (x)
If you should lose weight, then you should eat less.
Proper values and some facts about the world are sufficient to determine proper behavior.
What are proper values? Well, they’re the kind of values that determine proper behavior.
x: Saying this requirems me to know a moral fact. This moral fact is a consequence of an assumption I made about the true nature of reality. But to assume is to stoop lower than to define.
This is WillSawin_Should. NormalAnomaly_Should says the same thing, because we’re both humans. #$%^$_Should, where #$%^$ is the name of an alien from planet Mog, may say something completely different. You and I both use the letter sequence s-h-o-u-l-d to refer to the output of our own unique should-functions.
Lukeprog, the above is how I understand your post. Is it correct?
No. We both use the letter sequence “should” to direct our actions.
We believe that we should follow the results of our should-functions. We believe that the alien from Mog is wrong to follow the results of his should-function. These are beliefs, not definitions.
Imagine if you said “The sun will rise tomorrow” and I responded:
“This is NormalAnomaly_Will. WillSawin_Will says the same thing, because we’re both humans. #$%^$_Will, where #$%^$ is the name of an alien from planet Mog, may say something completely different. You and I both use the letter sequence w-i-l-l to refer to the output of our own unique will-functions.”
Normal_Anomaly’s ontology is coherent. What you describe regarding beliefs is also coherent but refers to a different part of reality space than what Normal is trying to describe.
I don’t understand what “ontology” and “reality space” mean in this context.
Here’s a guess:
You’re saying that the word “WillSawin_Should” is a reasonable word to use. It is well-defined, and useful in some contexts. But Plain-Old-Should is also a word with a meaning that is useful in some contexts.
in which case I would agree with you.
I was trying to convey that when you speak of beliefs and determination of actions you are describing an entirely different concept than what Normal_Anomaly was describing. To the extent that presenting your statements as a contradiction of Normal’s is both a conversational and epistemic error.
You can write_underscored_names by escaping the _ by preceding it with a \.
So you’re defining “should” to describe actions that best further one’s terminal values? Or is there an additional “shouldness” about terminal values too?
Also, regarding
in the grandparent, it sounds like you’re equivocating between defining what the word “ought” means and changing the true nature of the concept that “ought” usually refers to. (Unless I was wrong to add the “proper” in the quote, in which case I actually don’t know what point you were making.) To wit: “ought” is just a word that we can define as we like, but the concept that “ought” usually refers to is an adaptation and declaring that “ought” actually means something different will not change our actual behavior, except insofar as you succeed in changing others’ terminal values.
Incidentally this is a very slippery topic for me to talk about for reasons that I don’t fully understand, but I suspect it has to do with my moral intuitions constantly intervening and loudly claiming “no, it should be this way!” and the like. I also strongly suspect that this difficulty is nearly universal among humans.
There is.
You weren’t.
I do not think I am equivocating. Rather, I disagree with lukeprog about what people are changing when they disagree about morality.
lukeprog thinks that people disagree about what ought means / the definition of ought.
I believe that (almost) everybody things “ought” means the same thing, and that people disagree about the concept that “ought” usually refers to.
This concept is special because it has a reverse definition. Normally a word is defined by the situations in which you can infer that a statement about that word is true. However, “ought” is defined the other way—by what you can do when you infer that a statement about “ought” is true.
Is it the case that Katy ought to buy a car? Well, I don’t know. But I know that if Katy is rational, and she becomes convinced that she ought to buy a car, then she will buy a car.
What is the difference between what “ought” means and what it refers to?
Edit:
In the above, do you say that “You ought to do X.” is exactly equivalent to the command”Do X!”, and “I ought to do X.” means “I will do X on the first opportunity and not by accident.” ?
Ought we base the definition of “ought” on a pretty complicated notion of rationality?
To the first one, yes, but they have different connotations.
To the second one, sort of. “I” can get fuzzy here. I have akrasia problems. I should do my work, but I will not do it for a while. If you cut out a sufficiently small portion of my mind then this portion doesn’t have the opportunity to do my work until it actually does my work, because the rest of my mind is preventing it.
Furthermore I am thinking about them more internally. “should” isn’t part of predicting actions, its part of choosing them.
It doesn’t seem complicated to me. Certainly simpler than lukeprog’s definitions.
These issues are ones that should be cleared up by the discussion post I’m going to write in a second.
It seems that my further questions rather ought to wait a second, then.
It isn’t equivalent to a moral “ought”, since one person can command another to do something they both think is immoral.
This would require one of two situations:
a. A person consisting of multiple competing subagents, where the “ought” used by one is not the same as the “ought” used by another.
b. .A person with two different systems of morality, one dictating what is moral and the other how much they will accept deviating from it.
In either case you would need two words because there are two different kinds of should in the mind.
I gave the situation of one person commanding another. You replied with a scenario about one person with different internal systems. I don’t know why you did that.
It’s generally believed that if you shouldn’t tell people to do things they shouldn’t do.
So your problem reduces to the problem of someone who does things that they believe they shouldn’t.
If you’re not willing to make that reduction, I’ll have to think about things further.
I think it is obvious that involves someone doing something they think they shouldn’t. Which is not uncommon.
Which requires either a or b.
(yay, I finally caused a confusion that should be really easy to clear up!)
Alice and Bob agree that “Earth” means “that giant thing under us”. Alice and Bob disagree about the Earth, though. They disagree about that giant thing under them. Alice thinks it’s round, and Bob thinks it’s flat.
Yes, this is the distinction I had in mind.
So do you now think that I do not equivocate?
No, I think there is still equivocation in the claim that your dialog and Luke’s contradict one another. Luke is talking about the meaning of the word “Earth” and you are talking about the giant thing under us.
I also do not completely buy the assertion that “ought” is special because it has a reverse definition. This assertion itself sounds to me like a top-down definition of the ordinary type, if an unusually complex one.
Well there are two possible definitions, Luke’s and my reverse definition (or top-down definition of the ordinary type).
If you accept both definitions, then you have just proved that the right thing to do is XYZ. One shouldn’t be able to prove this just from definitions. Therefore you cannot accept both definitions.
Let’s try an analogy in another normative arena.
Suppose we propose to define rationality extensionally. Scientists study rationality for many decades and finally come up with a comprehensive definition of rationality that becomes the consensus. And then they start using that definition to shape their own inference patterns. “The rational thing to do is XYZ,” they conclude, using their definitions.
Where’s the problem?
Because the rational thing means the thing that produces good results, so you need a shouldness claim to scientifically study rationality claims, and science cannot produce shouldness claims.
The hidden assumption is something like “Good results are those produced by thinking processes that people think are rational” or something along those lines. If you accept that assumption, or any similar assumption, such a study is valid.
Let’s temporarily segregate rational thought vs. rational action, though they will ultimately need to be reconciled. I think that we can, and must, characterize rational thought first. We must, because “good results” are good only insofar as they are desired by a rational agent. We can, because while human beings aren’t very good individually at explicitly defining rationality, they are good, collectively via the scientific enterprise, at knowing it when they see it.
In this context, “science cannot produce shouldness claims” is contentious. Best not to make a premise out of it.
But why are rational thoughts good? Because they produce rational actions.
The circular bootstrapping way of doing this may not be untenable. (This is similar to arguments of Eliezer’s.)
You’re making hidden assumptions about what “good” and “rational” mean. Of course, some people accept those assumptions, that’s their prerogative. But those assumptions are not true by definition.
Yes I am making assumptions (they’re pretty open) about goodness and rationality. Or better: hypotheses. I advance them because I think theories constructed around them will ultimately cohere better with our most confident judgments and discoveries. Try them and see. By all means try alternatives and compare.
Circular bootstrapping is fine. And sure, the main reason rational thoughts are good is that they typically lead to good actions. But they’re still rational thoughts—thus correct—even when they don’t.
I think if you accept that you are making “assumptions” or “hypotheses” you agree with me.
Because you are thinking about the moral issue in a way reminiscent of scientific issues, as a quest for truth, not as a proof-by-definition.
I have difficulty to apply the analogy to ought.
I’m not really sure why this was downvoted, compared to everything else I’ve written on the topic.
Did it have to do with the excessive bolding somehow? Do my claims sound especially stupid stated like this?
It seems to completely miss Normal_Anomaly’s point, speaking right past him. As to the ‘compared to everything else you have written’ I refrained from downvoting your replies to myself even though I would have downvoted them if they were replies to a third party. It is a general policy of mine that I find practical, all else being equal.
Not for objective metaethicists, who seem to be able to escape your circle.
This doesn’t seem to actually be a term, after a few seconds of googling. Could you provide a link to a description of this philosophy?
I would much prefer to keep Luke’s. Basically because it is is actually useful when communicating with others who aren’t interested in having the other person’s values rammed down their throat. Because if you went around saying an ought at me using your definition then obviously you should expect me to reject it regardless of what the content is. Because the way you are using the term is such that it assumes that the recipient is ultimately subject to something that refers to your own mind.
So if you tell me I should go do something, and I agree with you, and I never go do that, you would see nothing inconsistent?
I’m totally comfortable with claims of the form “If you believe XYZ normative statements, then you should do W.” It should work just as well as conditionals about physical statements.
No, that is not something that is implied by my statements.
It is an example of someone not acting according to their own professed ideals and is inconsistent in the same way that all such things are.
So you’re saying that I am only allowed to use “should” to mean “WillSawin_should”. I can’t use it to mean “wedrifid_should”.
This seems like an odd way to run a conversation to me.
No, that is another rather bizarre thing which I definitely did not say. Perhaps it will be best for me to just leave it with my initial affirmation of Luke’s post:
In my observation Luke’s system for reducing moral claims provides more potential for enabling effective communication between agents and a more comprehensive way to form a useful epistemic model of such conversations.
So suppose I say:
“I wedrifid_should do X” and then don’t do X. Clearly, I am not being inconsistent.
but if I say:
“I should do X” and then don’t do X then I am being inconsistent.
Something must therefore prevent me from using “should” to mean “wedrifid_should”.
I’d agree that you can (and probably do) use plain old “should” to mean multiple things. The trouble is that this isn’t very useful for communication. So when communicating, us humans use heuristics to figure out what “should” is meant.
In the example of the conversation, if I say “you should X” and you say “I agree,” then I generally use a shortcut to think you meant Will-should. The obvious reason for this is that if you meant Manfred-should, you would have just repeated my own statement back to me, which would be not communicating anything, and it’s a decent shortcut to assume that when people say something they want to communicate. The only other obvious “should” in the conversation is Will-should, so it’s a good guess that you meant Will-should.
“I agree” generally means the same thing as repeating someone’s statement back at them. We can expand:
“You wedrifd_should do X”
“I agree, I will_should do X”
I seem to be making an error of interpretation here if I say things the way you normally say them! Why, in this instance, is it considered normal and acceptable to interpret professed agreement as expressing a different belief than the one being agreed to?
It all seems fishy to me.
Huh, yeah that is weird. But on thinking about it, I can only think of two situations I’ve heard or used “I agree.” One is if there’s a problem with an unsure solution, where it means “My solution-finding algorithm also returned that,” and if someone offers a suggestion about what should be done, where I seem to be claiming it usually means “My should-finding algorithm also returned that.”
In the first case, would you say that the \Manfred_solution is something or other? You and I mean something different by “solution”?
Of course not.
So why would you do something different for “should”?
Because there’s no objective standard against which “should algorithms” can be tested, like there is for the standard for “solution-finding algorithms” If there was no objective standard for solutions, I would absolutely stop talking about “the solution” and start talking about the Manfred_solution.
Didn’t you say in the other thread that we can disagree about the proper state of the world?
When we do that, what thing are we disagreeing about? It’s certainly not a standard, but how can it be subjective?
That’s the objective thing I am talking about.
Hm. I agree that you can disagree about some world-state that you’d like, but I don’t understand how we could move that from “we disagree” to “there is one specific world-state that is the standard.” So I stand by “no objective standard” for now.
I assume you are talking about proper or desirable world-states rather than actual ones.
I didn’t say it was the standard.
The idea is this.
If we disagree about what world state is best, there has to be some kind of statement I believe and you don’t, right? Otherwise, we wouldn’t disagree. Some kind of statement like “This world state is best.”
But the difference isn’t about some measurable property of the world, but about internal algorithms for deciding what to do.
Sure, to the extent that humans are irrational and can pit one desire against another, arguing about how to determine “best” is not a total waste of time, but I don’t think that has much bearing on subjectivity.
I’m losing the thread of the conversation at this point.
I have no solution to that problem.
Perhaps the meaning of the paragraph you quote wasn’t clear—I was trying hard to be polite rather than frank. You seem to be attacking a straw man using rhetorical questions so trivial that I would consider them disingenuous prior to adjusting for things like the illusion of transparency. Your conversation with lukeprog seems like one with more potential for useful communication. He cares about the subject far more than I do.
Moral ideas don’t determinne behaviour with any great reliability, so there is no analytical or necessary relationship there. If that’s what you were getting at.
I don’t want to spam but if people haven’t noticed then hopefully this comment should inform them that my first-ever lesswrong post, which might or might not make this clearer, is up.
I have created anchors for the sections, BTW:
Many Moral Reductionisms
Cognitivism vs. Non-Cognitivism
Objective vs. Subjective
Relative vs. Absolute
Moore’s Open Question Argument
The Is-Ought Gap
Moral Realism vs. Anti-Realism
Empathic Metaethics
(Replying to the comment instead of the post to be sure you’ll see this)
In “The Is-Ought Gap” you conclude
I have a lot to say about this particular issue but I’m not sure if you think you’ve exhausted the issue in this post or if you plan to come back to it. Just to begin with, though, I hope you’re aware that the two reasonable camps that take the gap most seriously would both agree with this conclusion. The issue is exactly this: we don’t think ‘ought’ refers.
I think you are incorrect with regards to Hume’s is-ought gap, although I find its relevance to be somewhat overstated. A hypothetical imperative such as your example relies on an equivocation between ‘ought’ as (1) a normative injunction and (2) conveying a possible causal pathway from here to there.
-
Here is the incorrect syllogism:
Premise 1: A desires C (is)
Premise 2: B will produce C (is)
Conclusion: A ought to do B (ought)
-
There is a hidden normative premise that is often ignored. It is
Premise 3: A should obtain its desires. (ought)
-
The correct syllogism would then be:
Premise 1 (is): A desires C
Premise 2 (is): B will produce C
Premise 3 (ought): A ought to obtain its desires.
Conclusion: A ought to do B (ought)
-
The necessity of Premise 3 is made clear by use of an admittedly extreme example:
P1: Hitler wants to kill a great number of people
P2: Zyklon B will kill a great number of people
C1: Hitler ought to use Zyklon B to kill a great number of people
While the conclusion is derived from the premises using definition (2) of the word ‘ought’, few would express it as a normative recommendation.
-
Hume’s fact/value dichotomy remains valid. A normative conclusion can only be validly deduced from a group of premises including at least one which is itself normative.
I don’t think it is impossible, but it is harder than you are making out. The examples given are not complete syllogisms, or other logical forms. It is easy to validly derive an ought form an is: you start with the factual statement and then invoke a bridging principle of the form:
if then
However, the argument is not sound unless the bridging statement is true. But the bridging statement is itself an derivation of an ought from an is, so there is a kind of circularity there. You are assuming that the ought-from-is problem has been solved in order to solve it.
As I said, I don’t think the situation is hopeless. The bridging premise is not exactly the same thing as a moral argument: it is usually more of a general statement along the lines of “if X increases well being, it should be done”. That provides some scope for an analytical justification of bridging principles.
[Re-post with correction]
Hi Luke,
I’ve questioned your metaethical views before (in your “desirist” days) and I think you’re making similar mistakes now as then. But rather than rehash old criticisms I’d like to make a different point.
Since you claim to be taking a scientific or naturalized approach to philosophy I would expect you to offer evidence in support of your position. Yet I see nothing here specifically identified as evidence, and very little that could be construed as evidence. I don’t see how your approach here is significantly different from the intuition-based philosophical approaches that you’ve criticised elsewhere.
How do you know this? Where’s the evidence? I don’t doubt that some people say, “Stealing is wrong because it’s against the will of God”. But where’s the evidence that they use “Stealing is wrong” to mean “Stealing is against the will of God”?
How do you know? And this seems to contradict your claim above that some people use “Stealing is wrong” to mean “stealing is against the will of God”. That’s not about what we want. (I say that moral terms are primarily about obligations, not wants.)
You can edit (pencil on paper icon) your posts, you don’t have to delete and repost.
Hi Luke,
I’ve questioned your metaethical views before (in your “desirist” days) and I think you’re making similar mistakes now as then. But rather than rehash old criticisms I’d like to make a different point.
Since you claim to be taking a scientific or naturalized approach to philosophy I would expect you to offer evidence in support of your position. Yet I see nothing here specifically identified as evidence, and very little that could be construed as evidence. I don’t see how your approach here is significantly different from the intuition-based philosophical approaches that you’ve criticised elsewhere.
How do you know this? Where’s the evidence? I don’t doubt that some people say, “Stealing is wrong because it’s against the will of God”. But where’s the evidence that they use “Stealing is wrong” to mean “Stealing is against the will of God”? (Indeed, if they meant that it would be very strange to say “Stealing is wrong because it’s against the will of God”. That would be equivalent to saying “A is true because A is true.” Yet it seems perfectly natural for someone to say this.)
How do you know? And this seems to contradict your claim above that some people use “Stealing is wrong” to mean “stealing is against the will of God”. That’s not about what we want. (I say that moral terms are primarily about obligations, not wants.)
That is of course, not what is right, but what she thinks is right. So far, so subjective.
Again, that is not the right thing, that is just what she thinks. An Objective metaethicist could answer the question what is right.
No: they are value terms about what we should want and be and do.
And the “we” is important here. Your metaethicists are like therapsists or life coaches or personal shoppers who advise people how to make their individual lives spiffier. But moral action is not solipsistic: moral choices affect other people. That’s why we can’t stop at “whatever you think is right is right”. I don’t want one of your metaethicists telling my neighbour how to be a better serial killer.
I think you’re missing the point of why there isn’t a universal code of ethics. If two people disagree about the definition of a word, the way forward isn’t to jump into the argument on one side or the other. The way forward is to stop using that word that way. This is subjective with respect to the word (if you don’t also specify how you’re defining it) but we stopped using that word that way anyhow—it’s objective with respect to the world, and that’s what’s important to how people act.
If someone uses “cat” to mean “animal that barks”, should everyone then stop using “cat”?
I can’t make any sense of that.
You’re right, it’s more complicated. It seems like the solution here is to make word choice a coordination problem, communication being a major goal of language—if a million people use it one way and one person uses it the other way, the one should say “an animal that barks.” On the other hand if everyone has the same several definitions for a word, like “sound,” then splitting up the word when necessary improves communication.
You complain that letting people specify what they mean by “right” makes “right” subjective where people diverge. But this doesn’t make the communication subjective if people replace “right” by an objective criterion for the world, so the bad stuff associated with just drifting off into subjectivity doesn’t happen.
Although I guess you could be saying that “right” being subjective is inherently bad. But I would suggest that you’re thinking about right_Peter, which is still objective.
If everyone has their own notion of right, we still have the Bad Thing that an action can only be allowed of forbidden, not Peter-allowed and Manfred-forbidden.
? So it’s impossible for two people to rationally disagree about whether or not an action is forbidden if the external state of the world is the same? I see no reason why “forbidden” in the moral sense should be objective.
If they disagree and they are both being rational, where’s the objectivity?
Try explaining to someone that something they like should be forbidden because you don’t like it.
I agree, it doesn’t look like there is much in this concept.
Okay. “If you don’t stop, I will shoot you.”
But seriously, WTF? Is that supposed to be an argument that if something is morally forbidden to one person it should be the same for another person?
If you don’t go looking for it, you won’t find it. As is so ofen the case on LW, that door has been shut without even trying to see what is behind it.
Arbitrary rules enforced with threats of violence is not an optimal outcome for me.
If you have an option other than
a) subjective laissez faire where serial killers are allowed to do their own thing
or
b) Tyranny
I’d be glad to hear it. I know I have.
Of course. No one should murder. I’m surprised you find that surprising.
Rather than chastising me, why not explain how “forbidden” is objective?
Re: shooting people: it was a joke. The WTF was not with respect to shooting people. It was because your demand was a non sequitur.
For start: are you aware of various (googleablle) standard defenses of metaethical ojectivism?
Is there some other place where you have defended metaethical (or even moral) objectivism?
I am aware of various defenses and find them all thoroughly lacking, but I doubt I have exhausted even the common possibilities. Go for it.
Assuming you’re generalized that properly and aren’t seriously arguing the most egregious false dichotomy I’ve seen in weeks, I’m afraid that condemning the set of ethics based on social or personal consequences as “tyranny” amounts to dismissing an entire school of thought on aesthetic grounds. Forgive me if I don’t find such a thing particularly convincing.
As I stated: “I know I have” I don’t think a) and b) are the only options either. I don’t see why the decision to call force-based ethics “tryrrany” counts as “aesthetic”. I don’t see why you are so hostile to reason-based ethics. You and various other people seem to think the rational/ojeciive approach to ethics needs to be dispensed with, but you don’t say why.
In conversations with that particular person, assuming they can’t easily be persuaded to change their usage? Yes, definitely.
That’s hardly an optimal outcome. They are making a mistake, although it seems no one wants to admit that.
Obviously the “optimal outcome” would be the easy persuasion I mentioned. Do you think someone misusing that word justifies arbitrary high effort in persuasion, or drastic measures?
Why do you think it’s the definition of the word that’s at issue?
Because it is possible for people to disagree about whether something is right or wrong without disagreeing about the state of the world.
Why must all disagreements be disagreements about the state of the world?
It seems to me like there are two kinds of disagreements, positive disagreements about the state of the world, and normative disagreements about the proper state of the world.
Nothing blows up when I believe that.
Knowing the fact that you just stated does not make fighting wars over morality seem less reasonable. In fact, it makes them seem more reasonable. Do you disagree? Do you think it makes sense to fight wars over a definition?
Sure. But if people start arguing over what’s right, they should argue over the proper state of the world, not over what’s “right.”
I don’t see much difference between “right” and “proper”.
He could also mean that we have to argue about states of the world.
But what else would we argue about, normatively? Abstract concepts, say, like “drugs are bad”. But then I would agree with him.
So I think we agree.
Again, I don’t follow you.
So perhaps he is saying that people should argue over the proper state of the world and not over the right XYZ, for some concept XYZ.
For instance, people should argue over the proper state of the world, not the right flavor of ice cream.
That is a true statement, there is certainly no objective right flavor of ice cream.
It is the most reasonable explanation I can come up with.
I agree that Luke’s approach has at times seemed implausibly individualistic. Moral reasoning is interpersonal from the get-go.
You’re missing the point. The empathic metaethicist is trying to figure out what she means by ‘right’. Assuming she’s a well-adjusted human being, that’s probably the same as what you mean by ‘right’, so with any luck we’ll work out what you mean by ‘right’ as well (and hence, what “is right”). But we’re not asking Alex what she thinks
peterdjones.getMeaning("right").getExtension()
is.That isn’t a good theoretical argument that “right” has only a subjective definition, and it isn’t practically as good as being able make individual notions of moral rightness more correct, where they need fixing.
Whatever you mean by “only a subjective definition”, I’m probably not trying to argue that.
Do you think you mean something other than what is right when you say “right”? If not, then replace “Alex” with “Peterdjones”. Do you still think the empathic metaethicist is going to get the wrong answer when they try to figure out what you mean by “right”?
Although I think this series of posts is interesting and mostly very well reasoned, I find the discussion about objectivity to be strangely crafted. At the risk of arguing about definitions: the hierarchy you lay out about objectivity is only remotely related to what I mean by objective, and my sense is that it doesn’t cohere very well with common usage.
First, there seems no better reason to split off objective1 than objectiveA which is “software-independent facts”. Okay, so I can’t say anything objective about my web browser, just because we’ve said I can’t. Why is this helpful? The only reason to split this out is if you are some sort of dualist; otherwise the mind is a computational phenomenon just like DNA replication or whatnot.
Second, as Emile already pointed out, nowhere in the hierarchy is uniqueness addressed, yet this is the clearest conventional distinction between subjectivity and objectivity. 5+7 = 12 for everyone. Mint chocolate chip ice cream is better than rocky road ice cream is not the case for everyone (in the conventional sense, anyway). So these things are all colloquially objective:
Rocky road has more chocolate than mint chocolate chip
The author of this post enjoys mint chocolate chip more than rocky road
My IPv4 address has a higher value than does lesswrong.org
The Bible describes God endorsing the consumption of only certain animals
Referring to God doesn’t make things non-objective in the standard sense presuming God exists. Of course, without a way to measure God’s preferences, you may lose your theoretical objectivity, but any other single source or self-consistent group can fill in (e.g. the Pope) as an source for objective answers to what would otherwise be subjective questions.
The issue isn’t whether that is subjective or objective; it’s whether that method of gaining objectivity is practical and useful.
And since humans are the only sentient beings, I really fail to see what the distinction is between 2 and 3 is in a practical way, once you split off God (or any other singularly identifiable entity).
So I strongly suggest that this section ought to be rethought. Objectivity seems central to this sort of moral reductionism, and so it is worth using definitions that are not too misleading. Either the definitions should change, or there should be much more motivation about why we care about the distinctions between any of the definitions you’ve offered.
In jest, I’m going to accuse you of plagiarizing my work, then tell you two problems that I have with the approach that you’ve outlined, and then wax e-peen and say that mine is similar, but more instructive on moral discourse among all users of it.
My problem here is that we already have a common language (in terms of wants) which reduces “should” and provides the kind of plurality that you’re seeking out of this approach, so there’s no need to claim, “I’m using ‘is good’ to mean P,” and then eek out a true statement whose truth is a matter of lexical elaboration, when instead people use moral language to alter others’ of their own behavior, perspective, etc. without all of that theorizing on top of it. Most people don’t make moral arguments on the basis of grand (meta)ethical stances. But it seems like everyone would have to be a deep ethicist to get any traction out of this theory, and that would mean that it really only explains how experts use moral language, but not how everyday people do. But why would one have to be deep about ethics to prescribe that someone do something? Can’t people prescribe an action and justify it without appealing to a definition of ‘goodness’ at all?
My last issue here is a potential paradox that I spotted when I made two pluralistic moral reductionists confront each other:
Are the words of this comment subjective3 (if we drop the specifier “moral” for a moment, given that this idea is not defined in this context)? They are determined by my reasoning, but they also obey the laws of physics. The notion of “independent” is not easy to make precise.
Right. It would take a lot of unpacking to make precise the term ‘mind-independent.’
I had similar thoughts when reading that passage.
There is (also) a certain sense in which the concept ‘objectively subjective’ is relevant here.
This is a false claim, unfortunately. Bachelor is not merely an “unmarried man”, it is an “unmarried man who could’ve been married in his society” (as all the long-discussed things like “#My 5-year-old son is a bachelor” and “#The Pope is a bachelor” show). ETA: the part beginning with “who” is probably a presupposition rather than assertion (“The Pope is not a bachelor” is only felicitous if used as metalinguistic “The Pope cannot be described by the word ‘bachelor’”, not if used in the literal sense “The Pope is married although it is not allowed”).
This one is also not obviously true. We can ask what Sherlock Holmes would approve of despite the fact that he never existed (and I can imagine a morality that says “good is what Sherlock approves of”—a strange morality though it would be). Why can’t we take “an essentially just and loving God” as a similar literature character?
Did the next few posts Luke mentions would be about empathic metaethics ever get written? I don’t see them anywhere.
Now that I’m running MIRI, I’ll probably never write it, but here’s where I was headed.
I’m commenting on the post-change “Is-Ought” section. It seems to me that most of the examples given of “ought” reductions do not support the conclusion that “the is-ought gap can be bridged”, because the reductions are wrong. Anyone can propose a naturalistic definition of “ought”, but at a minimum, to be right a translation of an “ought” statement into an “is” statement has to preserve the truth value of the “ought” statement, and most of the reductions listed fail to do so.
Take the first example:
If you give me any specific proposal for Y and Z, I’m pretty sure I can find an X such that “you ought to X” is obviously false and “X is obligatory (by deontic logic) if you assume axiomatic imperatives Y and Z” is true, or vice versa.
Luke, what is the meta-purpose of this sequence? It seems that you are trying to dissolve the persisting confusion about ethics, but what is the underlying purpose? I doubt this sequence is a terminal goal of yours, am I wrong? The reason for my doubt is that, before you started writing this sequence, you decided to join the visiting fellows program at the Singularity Institute for Artificial Intelligence. Which makes me infer that the purpose of this sequence has to do with friendly AI research.
If I am right that the whole purpose of this sequence has to do with friendly AI research, then how is it useful to devote so much resources to explaining the basics, instead of trying to figure out how to design, or define mathematically, something that could extrapolate human volition?
Luke confirms your guess in the comments section of this post.
I can see a couple of benefits of going over the basics first. One, it lets Luke confirm that his understanding of the basics are correct, and two, it can interest others to work in the same area (or if they are already interested, can help bring them up to the same level as Luke). I would like to see Luke continue this sequence. (Unless of course he already has some good new ideas about FAI, in which case write those down first!)
But how does ethics matter for friendly AI? If a friendly AI is going to figure out what humans desire, by extrapolating their volition, might it conclude that our volition is immoral and therefore undesirable?
What morality unit would it have other than human’s volition?
If it has another, separate volition unit, yes.
If not, then only if humans fundamentally self-contradict, which seems unlikely, because biological systems are pretty robust to that.
I am not sure what a ‘morality unit’ is supposed to be or how it would be different from a volition unit. Either morality is part of our volition, instrumental or an imperative. In each case one could ask what we want and arrive at morality.
What I’m saying is that: If Clippy tried to calculate our volition, he would conclude that our volition is immoral. (Probably. Maybe our volition IS paperclips.)
But if we programmed an AI to calculate our volition and use that as its volition, and our morality as its morality, and so on, then it would not find our volition immoral unless we find our volition immoral, which seems unlikely.
An AI that was smarter than us might deduce that we were not applying the Deep Structure of our morality properly because of bias or limited intelligence. It might conclude that human morality requires humans to greatly reduce their numbers in order to lessen the impact on other species, for instance.
Correct!
Hmm, that doesn’t cover the way I understand “objective” and “subjective”—I see them as referring to whether the answer to a question varies from one person to another, i.e. whether they are function of the person speaking. “Is that play interesting?” is subjective4, “Does that play follow the three unities of classical drama?” is objective4 - so to match your pattern it would be something like “Moral facts are objective4 if they are made true or false by facts independent of any single human, otherwise they are subjective4”.
i.e. it seems to me that some things are considered “objective” if they are merely social customs that may not hold in another society in another age.
But my thinking around this isn’t very clear, I’m mostly reacting to the fact that none of the definitions you listed seemed to fit the way I understood the words.
It seems to me that this subjective/objective difference can be somewhat dissolved by noticing (or, if you like, postulating) that “Is that play interesting?” is a different question when asked of Alice than when asked of Bob. The real question is, probably, in one case “Is that play interesting to Alice?” and in the other case “Is that play interesting to Bob?”.
In sentences like “Do you like vanilla ice cream?” it’s more explicit and clear that the meaning of “you” in it must be inferred from context (specifically, the identity of the questionee).
I would postulate that in general a “question” is directly about facts about the universe or mathematics (hopefully grounded in anticipation of sensory experience), but the sentence the question is asked in may contain ambiguous words which require context to interpret. And further, if the sentence together with the context is unambiguous, it indicates a “real question”.
In either case of the first example, the question has a single truth value which doesn’t change depending who you ask, though Alice and Bob may react differently to the sound waves (or light patterns) that are understood as “Is that play interesting?”.
Perhaps the sentences usually called “subjective4 questions” would be those containing such ambiguous words as “interesting” or “sexy”, in regard to which people are likely to suffer from the mind projection fallacy… I don’t know.
I don’t see why that would count as a dissolution. The subjective is defined as varying with individuals and the objective is defined as not so varying. All that your restatements as ”..”to Alice”, ”..to Bob” do is make that dependence explicit.
These words look ambiguous if you expect them to have a single value. They can be unambiguous if they have well defined values for different people. Subjective-ness isn’t ambiguity.
I guess I just don’t think it’s interesting or in any way special that there are questions that are usually asked by including non-verbal information, or that there are words that refer to “the recipient of this message”.
Wrt “ambiguity” all I mean is that you don’t know what the speaker intended the word to refer to until you know who they were talking to. Steve said “Is that play interesting?” to Alice, implying he wants to know whether Alice found the play interesting. Responding with Bob’s opinion of the play would be unhelpful, which is why you need the context. Maybe “ambiguous” isn’t the best word for that. Whatever.
That would be OK if a) there were a clear distinction between the two categories and/or b) nothing much rode on the distinction.
But neither is the case wrt morallity. a) We don’t have a “usual” practices with regard to moral language. Committed objectivists speak one way, subjectivists another, and many others are undecided b) it is hard to coneive of anything more important than morality—and the two ways of speaking mean something different. Alice can’t wish Bob to be punished just for doing something that’s wrong-for-Alice.
I agree that a very common meaning of objective/subjective is along your lines. In metaethics, it seems the most discussed. But I’d suggest function of the speaker’s beliefs and attitudes, not just function of the speaker. My left hand has five fingers: this truth is a function of the speaker, but if we want to communicate effectively in English, we’d better call this fact objective.
Answers can vary because people make mistakes about objective issues. You need to specify ideal agents to define objectivity, or to define subjectivity as answers that properly vary with individuals, ie the individual has the last word on their favourite flavour of ice cream.
Certainly, there are far more than 3 uses of the objective/subjective distinction! Check for footnote for a pointer to others.
Sounds like a form of speaker relativism, with the ‘empathetic’ project being about going beyond merely saying that people are expressing different fundamental standards, values, etc. to developing ways to bring those out into the open.
In a sense it can’t be considered robust realism. The two meanings of “sound” don’t lead to any confusion in practice: we know that powerful hi fis produce a lot of sound1 and we know that deaf people standing in front of the won’t hear any sound2.
However, morality is tied to the actions of oneself and others. There’s no coherent situation in which gay marriage can be both right and wrong, because there is no coherent way it can be both allowed and forbidden. In practice we tend to just average out preferences when it comes to defining law and allocating resources...but the preferences in question might as well be subjective. It is difficult to see what is being bought by the objectivity in pluralistic objective reductionism.
Of course normative pluralism doesn’t follow from descriptive pluralism. Some uses of moral terms be wrong. You said Craigs theological morality was wrong. If reducing and naturalising morality doesn’t allow you to say anybody is right or wrong, what is the point?
Or, eg,, not via God. Unless there are: if you can say 1 theory is wrong, you can say N-1 are wrong and arrive at one not-wrong theory by elimination. So I don’t see how your approach supports pluralism as a fixed principle: it can only depend on how things pan out in practice.
There’s also no coherent situation in which something can be sound and not sound.
Even worse than wrong (wrong here meaning wrong(Manfred) )! It follows from a probably false premise, and so is undefined when phrased as coming from God.
Ending the arguments over this stuff.
Sure there is: unheard compression waves.
Should we end all arguments by giving up on any fact of the matter?
So then let’s do the same for the gay marriage example.
We should end all arguments about subjects where there is no fact of the matter.
I don’t think you can. The different meanings of “sound” are disambiguated by context
OK. Then we just need to have the argument about whether there is a fact of the matter. Oh,,,we are.
Not anymore!
If “unheard compression waves” counts as an instance in which something is “sound and not sound”, then gay marriage can equally be “right and wrong” if, e.g., it is prohibited by the Bible but isn’t any obstacle to maximizing net preference satisfaction.
No doubt you want to protest: No, but there really is a single absolute truth about what’s right and what’s wrong, whereas there really isn’t a single absolute truth about what counts as “sound”. Maybe so, but that’s a highly disputable matter and you might do well to offer some actual arguments.
To convince people that it is ok for me to smack Urist McHatedRival in the face with a sharp rock and take all his stuff. And probably his daughters. The same as any other discussion about morality.
Are you telling me you are a Urist sympathiser and also practitioner of witchcraft? You ought not say such things. Come tribe, we should slay him too before the wrath of Odin falls upon us.
Moralizing is all about using hypocrisy effectively to achieve personal gains through social influence.
You haven’t argued that in any way. The point of my parable was that , while it may be easy to condemn other people’s uses of morality , it is much harder to do without when you feel vicitimised yourself.
Or you simply haven’t understood.
And your point would have be relevant if I was condemning the use of morality rather than describing them for the sake of giving a straight answer to a rhetorical question. Your ‘parable’ misses the mark. A “devil’s advocate” against a straw man, as it were.
Haven’t understood what? Please repeat whatever argument you believe I have missed.
Your comments about reality read like a contentious claim to me. Retroactively calling a contentious claim a “description” does nothing to remove the need for argumentative support.
This is still not making sense as a reply in the context. If you had expressed disagreement with a claim of mine then we could use arguments to try to persuade people about the subject. But if you present a refutation of something I didn’t say then it is an error for me to present arguments for whatever straw man you happened to to attack.
The following is the claim I do make:
One of the points of engaging in extensive debate about systems of morality is that while doing so you have the opportunity to influence the way your community thinks about how people should behave. This allows you to gain practical advantages for yourself and cause harm to your rivals.
I’m not going to provide an extended treatise on that subject here—it isn’t appropriate for the context. But if you do actually disagree with me then that gives you a clear position to argue against and I will leave you to do so without refutation.
I disagree your original claim that gaining advantage is the point of engaging in moral debate. Your revised claim, that it is only one of the points, is uninteresting, since any tool can be misused.
When people are talking about ‘ought’, they are frequently mean something that’s different from ‘is’ but is like ‘is’ in that it’s a primary concept. For them, ‘ought’ is not something that can be defined in terms of ‘is’.
So IMO people who are talking about ‘ought’ often really are talking about the world of ‘ought’, and that’s about all you can say about it.
You’re entitled to be uninterested in the world of ‘ought’ as a primary concept as well. I am not interested in it either, so I can’t defend the point of view of these ‘ought’ believers. I have repeatedly had conversations with them, so I am sure they exist.
Why stop with morality?For any debate about some X, you can shit that there is no real disagreement, and the two parties are just talking about two different things, X1 and X2, which you arrive at by treating one set of theoreticalclaims as the definition of X1 and the other as X2..
I don’t think this dissolution of disputes is a good idea in general, because I think theories aren’t definitions, and Ithink there are real disputes, and I’m suspicious of unuversal solvents. But I like tthe Sound example. But I dont like the morality example.. pluralism is bad for the sane reason that subhectivism is bad … morality has a practical, aspect… laws are passed, sanctions handed out and those either happen or don’t.
I’m confused. If this article promotes pluralistic moral reduction, why does Luke M. make a statement that sounds reductionaly singular rather than plural? I mean this statement here:
Isn’t this a reduction of morality to a single thing, specifically, “what we want”, i.e., desire?
Isn’t this defining morality as the practice of acting on our desires? And does not this definition contradict other definitions/reductions?
Am I correct in (roughly) summarizing your conclusion in the following quote?
If so, what is the logical difference between your theory and moral relativism? What if a person’s set of natural facts for morality is “those acts which the culture I was born into deem to be moral”?
Update: I’ve had the pleasure of discussing some of the topics from this post in episode 87 of video-podcast ‘Truth-Driven Thinking.’
I’m about to rewrite the section on the is-ought gap, for clarity, so here’s a copy of the original text:
Stealing often means wrongful taking of property… but point well taken.
True! Maybe I need a different example.
I miss the discussion (on LW in general) of an approach to ethics that strives to determine what actions should be unlawful for an agent, as opposed to, say, what probability distribution over actions is optimal for an agent. (And I don’t mean “deontologic”, as the “unlawfulness” can be predicated on the consequences.) If you criticize this comment for confusion of “descriptive ethics vs. normative ethics vs. metaethics”, try to be constructive.
Please explain why you think there should be more of that on LW.
What: discussion of the “social contract” aspect of ethics, for example of the right to not have one’s options (sets of actions) constrained beyond a threshold X, what that threshold should be, e.g. a property that the actions that infringe right-X of others are forbidden-X.
Why should there be more of that on LW: (1) it is equally practical and important aspect of ethics as the self-help aspect, (1) it seems to be simpler than determining optimal well-being conditions.
It would call for a series of ‘self help’ style of posts explaining:
The benefits of creating boundaries between your own identity and other people’s declarations of wrongness.
The art of balancing freedom with political expedience when dealing with other agents who are attempting to coerce you socially.
How to maintain internal awareness of the distinction between what you do not do for fear of social consequences vs what you do not do because of your own ethical values.
The difference between satisfying the preferences of others vs acquiescing to their demands. Included here would be how to deal with those who haven’t developed the ability to express their own desires except indirectly via the declarations of what it is ‘right’ for others do.
On the other hand, when i look at self help i see something i will continue to delay/slowly progress at a constant rate because my current situation seems to be quite similar to my ideal situation. I think that once you reach that point is when you start a more constant but passive process of improvement.
Your first #1 doesn’t seem to me to be a good justification for having more of it on LW. Lots of things are practical and important but don’t belong on LW.
Your second #1 seems to me wrong; deciding what’s actually right and wrong is very much not “simpler than determining optimal well-being conditions”, for the following reasons. (a) It’s debatable whether it’s even meaningful (since many people here are moral nonrealists or relativists of one sort or another). (b) There is no obvious way to reach agreement on what actually influences what’s right and what’s wrong. Net preference satisfaction? The will of a god? Obeying some set of ethical principles somehow built into the structure of the universe? Or what? (c) Most of the theories held by moral realists about what actually matters make it extraordinarily difficult to determine, in difficult cases, whether a given thing is right or wrong. Utilitarianism requires you to sum (or average, or something) the utilities of perhaps infinitely many beings, over a perhaps infinite extent of time and space. The theory Luke calls “desirism” requires you to work out the consequences of having many agents adopt any possible set of preferences. Intuitionist theories and divine-command theories make the details of what’s right and wrong entirely inaccessible. Etc.
Now, perhaps in fact you have some specific meta-ethical theory in mind such that, if that theory is true, then the ethical calculations become manageable. In that case, you might want to say what that meta-ethical theory is and why you think it makes the calculations manageable :-).
My answer to that question is that it what morality is actually about, and that personal preference-optimisation is something else.
I don’t think Luke, at least, is conflating morality with personal preference-optimization. He’s saying: Different people have different notions of “should”-ness, and if someone says “What should I do?” then giving them a good answer has to begin with working out what notion of “should” they’re working with. That applies whether “should” is being used morally or prudentially or both.
Also: What makes a moral agent a moral agent is having personal preferences that give substantial weight to moral considerations. And what such an agent is actually deciding, on any given occasion, is what serves his/her/its goals best: it’s just that among the important goals are things like “doing what is right” and “not doing what is wrong”. So, actually, for a moral agent “personal preference-optimization” will sometimes involve a great deal of “what morality is actually about”.
There’s an important difference between saying preferences may or may not include moral values, and saying morality is, by definition, preference-maximiisation.
Yup, there is. Did anyone say that morality is, by definition, preference-maximization?
Yes.
Do please feel free to provide more information.
I don’t criticize your comment on the basis of any confusion. It appears be more or less a coherent indication of preference. I criticize it based on considering the state which you desire to be both abhorrent and not (sufficiently) lacking here.
Do you find the “classification problem” variant of the “optimization problem” already repugnant, or is it something deeper?
Classification vs optimization is not necessarily a feature I was commenting on.
The degree of bullshit that is intrinsic to such conversations when engaged in by human participants may be a contributing factor.
This is the weakest part of the argument. Why should anybody believe that there is a super complicated function that determines what is ‘good’? What are the alternative hypotheses?
I can think of a much simpler hypothesis that explains all of the relevant facts. Our brains come equipped with a simple function that maps “is” statements to “ought” statements. Thus, we can reason about “ought” statements just like we do with “is” statements.
The special thing about this function is that there is nothing special about it at all. It is absolutely trivial. Any “ought” statement can potentially be inferred from any “is” statement. Therefore, “ought” statements can never be conditioned by evidence. This explains not only why there is lots of disagreement among people about what is “good” and that our beliefs about what constitutes “good” will be very complicated, but also that there will be no way to resolve these disagreements.
The situation is more complex than a set of objective moral truths that everyone agrees on, but it is also more complex than complete divergence. There is basis for convergence on the wrongness of murder and theft and some other things. Complete divergence would would mean itwould be irrational to even try to find common ground , and enshrine it in law.
I think the claim isn’t that there is a super-complicated function that determines what is ‘good’, but that the mapping from ‘is’ statements to ‘ought’ statements in the human brain is extremely complicated. If we claim that what is ‘good’ is what our brain considers is ‘good’, though, we merely encapsulate this complexity in a convenient black box.
That’s not to say that it’s not a solution, though: have you looked into desire utilitarianism? What you’re proposing here is really similar to (as I understand it) that school of moral philosophy claims. If you have time, Fyfe’s A Better Place is a good introduction.
The is-ought problem is an epistemic problem. Being informed that some A is ultimately, ontologically, the same as some B does not tell me how that A entails that B. If I cannot see how an “is” implies an “ought”, being informed that the “ought” ultimately refer to states of the world—states of the world far too complex for me to include in my epistemic calculations—does not help. I can’t cram a (representation of a) world-state into my brain. Being informed that if I could I would no longer have an is-ought problem under the unlikely circumstances that I could do so doesn’t help. The ontological claim that is’s and ought’s ultimately have the same referents can only be justified by some epistemic procedure. That is the only way any ontological claim is justified.
You really shouldn’t be using your own comments as evidence in an argument. It makes your reasoning appear… just a little motivated.
An argument works or it doens″t.
That’s true. Which means you really should have brought this argument up and resolved it, instead of making this argument and then declaring the matter unresolved.
I didn’t declared anything unresolved. I have argued that PMR does not close the is-ought gap. AFAIC that stands until someone counterargues. But hey, you could always downvote it out of visibiliy.