I’ve been working my way through the Sequences—and I’m wondering a lot about this essay, in light of the previously-introduce notion of ‘how do you decide what values, given to you by natural selection, you are going to keep?’
Could someone use the stances you develop here, EY, to argue for something like Aristotelian ethics? (Which, admittedly, I may not properly understand fully, but my basic idea is:)
‘You chose to keep human life, human happiness, love, and learning as values in YOUR utility function,’ says the objector, ’even though you know where they came from. You decided that you wanted them anyway. You did this because you had to start somewhere, and you claim that if you stripped away everything provided by natural selection you wouldn’t be left with anything. Under the same logic, why can’t I keep all the ethical injunctions as terminal values?
‘Your explanation of where ‘the ends does not justify the means’, is very clever and all. Your explanation of ‘thou shalt not kill’ is very clever. But so what if we know where they came from? If we know why nature selected on them, in our specific case? I’m no more obligated to dispose of it than I am to dispose of ‘human happiness is good’.′
Is the counter-argument simply that this leads to a utility function you would call inconsistent?
Oh, and...sorry for commenting on all these dead threads...it’s a pity I got here so late.
I understand why the notions exist—I was trying to address the question of ‘what explainable-moral-intuitions should we keep as terminal values, and how do we tell them apart from those we shouldn’t’.
But your first sentence is taken very much to heart, sir.
Maybe I’m being silly here, in hindsight. Certain intuitive desires are reducible to others, and some, like ‘love/happiness/fun/etc.’ are probably not. It feels obvious that most people should immediately see that. Yes, they want a given ethical injunction to be obeyed, but not as a fundamental/terminal value.
Then again—there are Catholic moralists, including, I think, some Catholics I know personally, who firmly believe that (for example) stealing is wrong because stealing is wrong. Not for any other reason. Not because it brings harm to the person being stolen from. If you bring up exceptions—‘what about an orphan who will starve if they don’t steal that bread?’ they argue that this doesn’t count as stealing, not that it ‘proves that stealing isn’t really wrong.’ For them, every exception is simply to be included as another fundamental rule. At least, that’s the mindset, as far as I can tell. I saw the specific argument above being formulated for use against moral relativists, who were apparently out to destroy society by showing that different things were right for different people.
Even though this article is about AI, and even though we should not trust ourselves to understand when we should be excepted from an injunction—this seems like a belief that might eventually have some negative real-world consequences. See, potentially, ‘homosexuality is wrong because homosexuality is wrong’?
If I tried to tell any of these people about how ethical injunctions could be explained as heuristics for achieving higher terminal values—I can already feel myself being accused of shuffling things around, trying to convert goods into other incompatible goods in order to justify some sinister, contradictory worldview.
If I brought up reductionism, it seems almost trivial—while I’m simulating their mind—to point out that no one has ever provably applied reduction to morals.
So maybe let me rephrase: is there any way I could talk them out of it?