Where does your belief regarding suffering come from?
Does it come from ordinary human values?
great, CEV will handle it.
Does it come from your own personal unique values?
the rest of humanity has no obligation to go along with that
Does it come from pure logic that the rest of us would realize if we were smart enough?
great, CEV will handle it.
Is it just a brute fact that suffering of all entities whatsoever is bad, regardless of anyone’s views? And furthermore, you have special insight into this, not from your own personal values, or from logic, but...from something else?
then how are you not a religion? where is it coming from?
It is not clear to me exactly what “belief regarding suffering” you are talking about, what you mean by “ordinary human values”/”your own personal unique values”.
As I argue in Section 2.2., there is (at least) a non-negligible chance that s-risks occur as a result of implementing human-CEV, even if s-risks are very morally undesirable (either in a realist or non-realist sense).
Please read the paper, and if you have any specific points of disagreement cite the passages you would like to discuss. Thank you
It is not clear to me exactly what “belief regarding suffering” you are talking about, what you mean by “ordinary human values”/”your own personal unique values”.
Belief regarding suffering: the belief that s-risks are bad, independently of human values as would be represented in CEV.
Ordinary human values: what most people have.
Your own personal unique values: what you have, but others don’t.
Please read the paper, and if you have any specific points of disagreement cite the passages you would like to discuss. Thank you
In my other reply comment, I pointed out disagreements with particular parts of the paper you cited in favour of your views. My fundamental disagreement though, is that you are fundamentally relying on an unjustified assumption, repeated in your comment above:
even if s-risks are very morally undesirable (either in a realist or non-realist sense)
The assumption being that s-risks are “very morally undesirable”, independently of human desires (represented in CEV).
You ask: “Where does your belief regarding the badness of s-risks come from?”
And you provide 3 possible answers I am (in your view) able to choose between:
“From what most people value” 2. “From what I personally value but others don’t” or 3. “from pure logic that the rest of us would realize if we were smart enough”.
However, the first two answers do not seem to be answers to the question. My beliefs about what is or is not morally desirable do not come from “what most people value” or “what I personally value but others don’t”. In one sense my beliefs about ethics, as everyone’s beliefs about ethics, come from various physical causes (personal experiences, conversations I have had with other people, papers I have read) such as in the formation of all other kinds of beliefs. There is another sense in which my beliefs about ethics, seem to me to be justified by reasons/preferences. This second sense, I believe is the one you are interested in discussing. And what is exactly the nature of the reasons or preferences that make me have certain ethical views is what the discipline of meta-ethics is about. To figure out or argue for which is the right position in meta-ethics is outside the scope of this paper, which is why I have not addressed it in the paper. Below I will reply to your other comment and discuss more the meta-ethical issue.
Suppose that my definition of “suffering” (as a moral term) was “suffering by a human” and my definition of “s-risk” was “risk of massive risk of suffering by humans”, and my definition of ‘human’ was a member of the biological species Homo sapiens (or a high-fidelity upload of one). You tell me we have to pay attention to animal suffering and animal s-risks, and I say “while the biological phenomenon of pain in humans and in animals is identical, in my ethical system human have moral weight and animals don’t. So animal pain is not, morally speaking, suffering, and risk of it is not s-risk.” You say “oh yes it is”, and I say “by your ethical systems axioms, yes, but not by mine”. How do you then persuade me otherwise, using only ethics and logic, when you and I don’t operate in the same ethical system? You’re just saying “I have axiom A”, and my reply is “good for you, I have axiom B”.
You can’t use logic here, because you and your interlocutor don’t share the same axiom system. However, you can say “A society that used my proposed ethical system would produce outcome X, whereas a society using yours would produce outcome Y, and pretty-much every human finds X cute and fluffy and Y nauseating, that’s just the way human instincts are. So even though all you care about is humans, my ethical system is better.” That’s a valid argument that might win, ethical logic isn’t. You have to appeal to instinct and/or aesthetics, because that’s all you and your interlocutor (hopefully) agree on.
Hi Roger, first, the paper is addressed to those who already do believe that all sentient beings deserve moral consideration and that their suffering is morally undesirable. I do not argue for these points in the paper, since they are already universally accepted in the moral philosophy literature.
This is why, for instance, write the following: “sentience in the sense understood above as the capacity of having positively or negatively valenced phenomenally conscious experiences is widely regarded and accepted as a sufficient condition for moral patienthood (Clarke, S., Zohny, H. & Savulescu, J., 2021)”.
Furthermore, it is just empirically not the case that people cannot be convinced “only by ethics and logic”: for instance, many people reading Peter Singer’s Animal Liberation, as a result, changed their views in light of the arguments he provided in the first chapter and came to believe that non-human animals deserve equal moral consideration of interests. Changing one’s ethical views when presented with ethical arguments is a standard practice that occurs to moral philosophers when researching and reading moral philosophy. Of course, there is the is/ought to gap, but this does not entail that one cannot convince someone that the most coherent version of their most fundamental ethical intuitions do not, in fact, lead where they believe they lead but instead that they lead to somewhere else, to a different conclusion. This happens all the time between more philosophers, one presents an argument in favour of a view, and in many instances, many philosophers are convinced by that argument and change their view.
In this paper, I was not trying to argue that non-human animals deserve moral consideration or that s-risks are bad, as I said, I have assumed this. What I try to argue is that if this is true, then, in some decision-making situations we would have some strong pro-tanto moral reasons to implement SCEV. In fact, I do not even argue that conclusively, what we should do is try to implement SCEV.
the paper is addressed to those who already do believe that all sentient beings deserve moral consideration and that their suffering is morally undesirable.
I think you should state these assumptions more clearly at the beginning of the paper, since you appear to be assuming what you are claiming to prove. You are also making incorrect assumptions about your audience, especially when posting it to Less Wrong. The idea that Coherent Extrapolated Volition, Utilitarianism, or “Human Values” applies only to humans, or perhaps only to sapient beings, is quite widespread on Less Wrong.
I do not argue for these points in the paper, since they are already universally accepted in the moral philosophy literature
I’m not deeply familiar with the most recent few decades of the moral philosophy literature, so I won’t attempt to argue this in a recent context, if that is what you in fact mean by “the moral philosophy literature” (though I have to say that I do find any claim of the form “absolutely everyone who matters agrees with me” inherently suspicious). However, Philosophy is not a field that has made such rapid recent advances such that one can simply ignore all but the last few decades, and for the moral philosophy literature of the early 20th century and the preceding few millennia (which includes basically every philosopher named in a typical introductory guide to Moral Philosophy), this claim is just blatantly false, even to someone from outside the academic specialty. For example, I am quite certain that Nietzsche, Hobbes, Thomas Aquinas and Plato would all have variously taken issue with the proposition that humans and ants deserve equal moral consideration, if ants can be shown to experience pain (though the Jains would not). Or perhaps you would care to cite quotes from each of them clearly supporting your position? Indeed, for much of the last two millennia, Christian moral philosophy made it entirely clear that they believed animals do not have souls, and thus did not deserve the same moral consideration as humans, and that humans held a unique role in God’s plan, as the only creature made in His image and imbued with souls. So claiming that your position is “already universally accepted in the moral philosophy literature” while simply ignoring O(90%) of that literature appears specious to me. Perhaps you should also briefly outline in your paper which portions of or schools from the moral philosophy literature in fact agree with your unstated underlying assumption?
What I mean by “moral philosophy literature” is the contemporary moral philosophy literature, I should have been more specific, my bad. And in contemporary philosophy, it is universally accepted (though of course, the might exist one philosopher or another who disagrees) that sentience in the sense understood above as the capacity of having positively or negatively valenced phenomenally conscious experiences is sufficient for moral patienthood. If this is the case, then, it is enough to cite a published work or works in which this is evident. This is why I cite Clarke, S., Zohny, H. & Savulescu, J., 2021. You can go see this recently edited book on moral status that this claim is assumed thought and in the book you can find the sources for its justification.
OK, note to self: If we manage to create a superintelligece, and give us access to the contemporary moral philosophy literature, it will euthanize us all and feed us to ants. Good to know!
I do not think this follows, the “consensus” is that sentience is sufficient for moral status. It is not clearly the case that giving some moral consideration to non-human sentient beings would lead to the scenario you describe. Though see: https://www.tandfonline.com/doi/full/10.1080/21550085.2023.2200724
“Some”, or “pro-tanto” unspecified amount of moral consideration, I agree in principle. “Equal” or even “anywhere within a few orders of magnitude of equal”, and we go extinct. Ants need ~10,000,000 times less resources per individual than humans, so if you don’t give humans around ~10,000,000 times the moral value, we end up extinct in favor of more ants. For even tinier creatures, the ratios are even larger. Explaining why moral weight ought to scale linearly with body weight over many orders of magnitude is a challenging moral position to argue for, but any position that doesn’t closely approximate that leads to wildly perverse incentives and the “repugnant conclusion”. The most plausible-sounding moral argument I’ve come up with is that moral weight should be assigned somewhat comparably per-species at a planetary level, and then shared out (equally?) per individual member of a species, so smaller more-numerous species end up with a smaller share per individual. However, given my attitude of ethical system design, I view these sorts of arguments as post-facto political-discussion justifications, and am happy to do what works, and between species of very different sizes, the only thing that works is that moral weight scales roughly linearly with adult body weight (or more accurately, resource needs).
I enjoyed Jeff Sebo’s paper, thank-you for the reference, and mostly agree with his analysis, if not his moral intuitions — but I really wish he had put some approximate numbers in on occasion to show just how many orders of magnitude the ratios can be between the “large” and “small” things he often discusses. Those words conjure up things within an order of magnitude of each other, not many orders-of-magnitude apart. Words like “vast” and “minute” might have been more appropriate, even before he got on to discussing microbes. But I loved Pascal’s Bugging.
Overall, thank-you for the inspiration: Due to your paper and this conversation, I’m now working on another post for my AI, Alignment and Ethics sequence where I’ll dig in more depth into this exact question, of the feasibility or otherwise of granting moral worth to sentient animals, from my non-moral-absolutist ethical-system design viewpoint, This one’s a really hard design problem that requires a lot of inelegant hacks. My urgent advice would be to steer clear of it, at least unless you have very capable ASI assistance and excellent nanotech and genetic engineering, plus some kind of backup plan in case you made a mistake and persuaded your ASIs to render humanity extinct. Something like an even more capable ASI running the previous moral system ready to step in under prespecified circumstances comes to mind, but then how do you get it to not step in due to ethical disagreement?.
I am glad to hear you enjoyed the paper and that our conversation has inspired you to work more on this issue! As I mentioned I now find the worries you lay out in the first paragraph significantly more pressing, thank you for pointing them out!
To the people downvoting/disagreeing, tell me:
Where does your belief regarding suffering come from?
Does it come from ordinary human values?
great, CEV will handle it.
Does it come from your own personal unique values?
the rest of humanity has no obligation to go along with that
Does it come from pure logic that the rest of us would realize if we were smart enough?
great, CEV will handle it.
Is it just a brute fact that suffering of all entities whatsoever is bad, regardless of anyone’s views? And furthermore, you have special insight into this, not from your own personal values, or from logic, but...from something else?
then how are you not a religion? where is it coming from?
It is not clear to me exactly what “belief regarding suffering” you are talking about, what you mean by “ordinary human values”/”your own personal unique values”.
As I argue in Section 2.2., there is (at least) a non-negligible chance that s-risks occur as a result of implementing human-CEV, even if s-risks are very morally undesirable (either in a realist or non-realist sense).
Please read the paper, and if you have any specific points of disagreement cite the passages you would like to discuss. Thank you
Belief regarding suffering: the belief that s-risks are bad, independently of human values as would be represented in CEV.
Ordinary human values: what most people have.
Your own personal unique values: what you have, but others don’t.
In my other reply comment, I pointed out disagreements with particular parts of the paper you cited in favour of your views. My fundamental disagreement though, is that you are fundamentally relying on an unjustified assumption, repeated in your comment above:
The assumption being that s-risks are “very morally undesirable”, independently of human desires (represented in CEV).
Okay, I understand better now.
You ask: “Where does your belief regarding the badness of s-risks come from?”
And you provide 3 possible answers I am (in your view) able to choose between:
“From what most people value” 2. “From what I personally value but others don’t” or 3. “from pure logic that the rest of us would realize if we were smart enough”.
However, the first two answers do not seem to be answers to the question. My beliefs about what is or is not morally desirable do not come from “what most people value” or “what I personally value but others don’t”. In one sense my beliefs about ethics, as everyone’s beliefs about ethics, come from various physical causes (personal experiences, conversations I have had with other people, papers I have read) such as in the formation of all other kinds of beliefs. There is another sense in which my beliefs about ethics, seem to me to be justified by reasons/preferences. This second sense, I believe is the one you are interested in discussing. And what is exactly the nature of the reasons or preferences that make me have certain ethical views is what the discipline of meta-ethics is about. To figure out or argue for which is the right position in meta-ethics is outside the scope of this paper, which is why I have not addressed it in the paper. Below I will reply to your other comment and discuss more the meta-ethical issue.
Suppose that my definition of “suffering” (as a moral term) was “suffering by a human” and my definition of “s-risk” was “risk of massive risk of suffering by humans”, and my definition of ‘human’ was a member of the biological species Homo sapiens (or a high-fidelity upload of one). You tell me we have to pay attention to animal suffering and animal s-risks, and I say “while the biological phenomenon of pain in humans and in animals is identical, in my ethical system human have moral weight and animals don’t. So animal pain is not, morally speaking, suffering, and risk of it is not s-risk.” You say “oh yes it is”, and I say “by your ethical systems axioms, yes, but not by mine”. How do you then persuade me otherwise, using only ethics and logic, when you and I don’t operate in the same ethical system? You’re just saying “I have axiom A”, and my reply is “good for you, I have axiom B”.
You can’t use logic here, because you and your interlocutor don’t share the same axiom system. However, you can say “A society that used my proposed ethical system would produce outcome X, whereas a society using yours would produce outcome Y, and pretty-much every human finds X cute and fluffy and Y nauseating, that’s just the way human instincts are. So even though all you care about is humans, my ethical system is better.” That’s a valid argument that might win, ethical logic isn’t. You have to appeal to instinct and/or aesthetics, because that’s all you and your interlocutor (hopefully) agree on.
Hi Roger, first, the paper is addressed to those who already do believe that all sentient beings deserve moral consideration and that their suffering is morally undesirable. I do not argue for these points in the paper, since they are already universally accepted in the moral philosophy literature.
This is why, for instance, write the following: “sentience in the sense understood above as the capacity of having positively or negatively valenced phenomenally conscious experiences is widely regarded and accepted as a sufficient condition for moral patienthood (Clarke, S., Zohny, H. & Savulescu, J., 2021)”.
Furthermore, it is just empirically not the case that people cannot be convinced “only by ethics and logic”: for instance, many people reading Peter Singer’s Animal Liberation, as a result, changed their views in light of the arguments he provided in the first chapter and came to believe that non-human animals deserve equal moral consideration of interests. Changing one’s ethical views when presented with ethical arguments is a standard practice that occurs to moral philosophers when researching and reading moral philosophy. Of course, there is the is/ought to gap, but this does not entail that one cannot convince someone that the most coherent version of their most fundamental ethical intuitions do not, in fact, lead where they believe they lead but instead that they lead to somewhere else, to a different conclusion. This happens all the time between more philosophers, one presents an argument in favour of a view, and in many instances, many philosophers are convinced by that argument and change their view.
In this paper, I was not trying to argue that non-human animals deserve moral consideration or that s-risks are bad, as I said, I have assumed this. What I try to argue is that if this is true, then, in some decision-making situations we would have some strong pro-tanto moral reasons to implement SCEV. In fact, I do not even argue that conclusively, what we should do is try to implement SCEV.
I think you should state these assumptions more clearly at the beginning of the paper, since you appear to be assuming what you are claiming to prove. You are also making incorrect assumptions about your audience, especially when posting it to Less Wrong. The idea that Coherent Extrapolated Volition, Utilitarianism, or “Human Values” applies only to humans, or perhaps only to sapient beings, is quite widespread on Less Wrong.
I’m not deeply familiar with the most recent few decades of the moral philosophy literature, so I won’t attempt to argue this in a recent context, if that is what you in fact mean by “the moral philosophy literature” (though I have to say that I do find any claim of the form “absolutely everyone who matters agrees with me” inherently suspicious). However, Philosophy is not a field that has made such rapid recent advances such that one can simply ignore all but the last few decades, and for the moral philosophy literature of the early 20th century and the preceding few millennia (which includes basically every philosopher named in a typical introductory guide to Moral Philosophy), this claim is just blatantly false, even to someone from outside the academic specialty. For example, I am quite certain that Nietzsche, Hobbes, Thomas Aquinas and Plato would all have variously taken issue with the proposition that humans and ants deserve equal moral consideration, if ants can be shown to experience pain (though the Jains would not). Or perhaps you would care to cite quotes from each of them clearly supporting your position? Indeed, for much of the last two millennia, Christian moral philosophy made it entirely clear that they believed animals do not have souls, and thus did not deserve the same moral consideration as humans, and that humans held a unique role in God’s plan, as the only creature made in His image and imbued with souls. So claiming that your position is “already universally accepted in the moral philosophy literature” while simply ignoring O(90%) of that literature appears specious to me. Perhaps you should also briefly outline in your paper which portions of or schools from the moral philosophy literature in fact agree with your unstated underlying assumption?
What I mean by “moral philosophy literature” is the contemporary moral philosophy literature, I should have been more specific, my bad. And in contemporary philosophy, it is universally accepted (though of course, the might exist one philosopher or another who disagrees) that sentience in the sense understood above as the capacity of having positively or negatively valenced phenomenally conscious experiences is sufficient for moral patienthood. If this is the case, then, it is enough to cite a published work or works in which this is evident. This is why I cite Clarke, S., Zohny, H. & Savulescu, J., 2021. You can go see this recently edited book on moral status that this claim is assumed thought and in the book you can find the sources for its justification.
OK, note to self: If we manage to create a superintelligece, and give us access to the contemporary moral philosophy literature, it will euthanize us all and feed us to ants. Good to know!
I do not think this follows, the “consensus” is that sentience is sufficient for moral status. It is not clearly the case that giving some moral consideration to non-human sentient beings would lead to the scenario you describe. Though see: https://www.tandfonline.com/doi/full/10.1080/21550085.2023.2200724
“Some”, or “pro-tanto” unspecified amount of moral consideration, I agree in principle. “Equal” or even “anywhere within a few orders of magnitude of equal”, and we go extinct. Ants need ~10,000,000 times less resources per individual than humans, so if you don’t give humans around ~10,000,000 times the moral value, we end up extinct in favor of more ants. For even tinier creatures, the ratios are even larger. Explaining why moral weight ought to scale linearly with body weight over many orders of magnitude is a challenging moral position to argue for, but any position that doesn’t closely approximate that leads to wildly perverse incentives and the “repugnant conclusion”. The most plausible-sounding moral argument I’ve come up with is that moral weight should be assigned somewhat comparably per-species at a planetary level, and then shared out (equally?) per individual member of a species, so smaller more-numerous species end up with a smaller share per individual. However, given my attitude of ethical system design, I view these sorts of arguments as post-facto political-discussion justifications, and am happy to do what works, and between species of very different sizes, the only thing that works is that moral weight scales roughly linearly with adult body weight (or more accurately, resource needs).
I enjoyed Jeff Sebo’s paper, thank-you for the reference, and mostly agree with his analysis, if not his moral intuitions — but I really wish he had put some approximate numbers in on occasion to show just how many orders of magnitude the ratios can be between the “large” and “small” things he often discusses. Those words conjure up things within an order of magnitude of each other, not many orders-of-magnitude apart. Words like “vast” and “minute” might have been more appropriate, even before he got on to discussing microbes. But I loved Pascal’s Bugging.
Overall, thank-you for the inspiration: Due to your paper and this conversation, I’m now working on another post for my AI, Alignment and Ethics sequence where I’ll dig in more depth into this exact question, of the feasibility or otherwise of granting moral worth to sentient animals, from my non-moral-absolutist ethical-system design viewpoint, This one’s a really hard design problem that requires a lot of inelegant hacks. My urgent advice would be to steer clear of it, at least unless you have very capable ASI assistance and excellent nanotech and genetic engineering, plus some kind of backup plan in case you made a mistake and persuaded your ASIs to render humanity extinct. Something like an even more capable ASI running the previous moral system ready to step in under prespecified circumstances comes to mind, but then how do you get it to not step in due to ethical disagreement?.
I am glad to hear you enjoyed the paper and that our conversation has inspired you to work more on this issue! As I mentioned I now find the worries you lay out in the first paragraph significantly more pressing, thank you for pointing them out!