RFC: Meta-ethical uncertainty in AGI alignment
I’m working on writing a paper about an idea I previously outlined for addressing false positives in AI alignment research. This is the first completed draft of one of the subsections arguing for the adoption of a particular, necessary hinge proposition to reason about aligned AGI. I appreciate feedback on this subsection especially regarding if you agree with the line of reasoning and if you think I’ve ignored anything important that should be addressed here. Thanks!
AGI alignment is typically phrased in terms of aligning AGI with human interests, but this hides some of the complexity of the problem behind determining what “human interests” are. Taking “interests” as a synonym for “values”, we can begin to make some progress by treating alignment as at least partially the problem of teaching AGI human values (Soares, 2016). Unfortunately, what constitutes human values is currently unknown since humans may not be aware of the extent of their own values or may not hold reflexively consistent values (Scanlon, 2003). Further complicating matters, humans are not rational, so their values cannot be deduced from their behavior unless some normative assumptions are made (Tversky, 1969), (Armstrong and Mindermann, 2017). This is a special case of Hume’s is-ought problem—that axiology cannot be inferred from ontology alone—and it complicates the problem of training AGI on human values (Hume, 1739).
Perhaps some of the difficulty could be circumvented if a few normative assumptions were made, like assuming that rational preferences are always better than irrational preferences or assuming that suffering supervenes on preference satisfaction. This poses an immediate problem for our false positive reduction strategy by introducing additional variables that will necessarily increase the chance of a false positive. Maybe we could avoid making any specific normative assumptions prior to the creation of aligned AGI by expecting the AGI to discover them via a process like Yudkowsky’s coherent extrapolated volition (Yudkowsky, 2004). This may avoid the need to make as many assumptions, but still requires making at least one—that moral facts exist to permit the correct choice of normative assumptions—and reveals a deep philosophical problem at the heart of AGI alignment—meta-ethical uncertainty.
Meta-ethical uncertainty stems from epistemic circularity and the problem of the criterion since it is not possible to know the criteria by which to asses which moral facts are true or even if any moral facts exist without first assuming to know what is good and true (Chisholm, 1982). We cannot hope to resolve meta-ethical uncertainty here, but we can at least decide what impact particular assumptions about the existence of moral facts have upon false positives in AGI alignment. Specifically, whether or not moral facts exists and, if they do, what moral facts should be assumed to be true.
On the one hand suppose we assume that moral facts exist, then we could build aligned AGI on the presupposition that it could at least discover moral facts even if no moral facts were specified in advance and then use knowledge of these facts to constrain its values such that they aligned with humanity’s values. Now suppose this assumption is false and moral facts do not exist, then our moral-facts-assuming AGI would either never discover any moral facts to constrain its values to be aligned with human values or would constrain itself with arbitrary moral facts that would not be sure to produce value alignment with humanity.
On the other hand suppose we assume that moral facts do not exist, then we must build aligned AGI to reason about and align itself with the axiology of humanity in the absence of any normative assumptions, likely on a non-cognitivist basis like emotivism. Now suppose this assumption is false and moral facts do exist, then our moral-facts-denying AGI would discover the existence of moral facts, at least implicitly, by their influence on the axiology of humanity and would align itself with humanity as if it had started out assuming moral facts existed but at the cost of solving the much harder problem of learning axiology without the use of normative assumptions.
Based on this analysis it seems that assuming the existence of moral facts, let alone assuming any particular moral facts, is more likely to produce false positives than assuming moral facts do not exist because denying the existence of moral facts gives up the pursuit of a class of alignment schemes that may fail, namely those that depend on the existence of moral facts. Doing so likely makes finding and implementing a successful alignment scheme harder, but it does this by replacing difficulty tied to uncertainty around a metaphysical question that may not be resolved in favor of alignment to uncertainty around implementation issues that through sufficient effort may be made to work. Barring a result showing that moral nihilism—the assumption that no moral facts exist—implies the impossibility of building aligned AGI, it seems the best hinge proposition to hold in order to reduce false positives in AGI alignment due to meta-ethical uncertainty.
References:
Nate Soares. The Value Learning Problem. In Ethics for Artificial Intelligence Workshop at 25th International Joint Conference on Artificial Intelligence. (2016). Link
T. M. Scanlon. 3 Rawls on Justification. 139 In The Cambridge Companion to Rawls. Cambridge University Press, 2003.
Amos Tversky. Intransitivity of preferences.. Psychological Review 76, 31–48 American Psychological Association (APA), 1969.Link
Stuart Armstrong, Sören Mindermann. Impossibility of deducing preferences and rationality from human policy. (2017). Link
David Hume. A Treatise of Human Nature. Oxford University Press, 1739. Link
Eliezer Yudkowsky. Coherent Extrapolated Volition. (2004). Link
Roderick M. Chisholm. The Foundations of Knowing. University of Minnesota Press, 1982.
- AI Alignment Problem: “Human Values” don’t Actually Exist by 22 Apr 2019 9:23 UTC; 45 points) (
- Value uncertainty by 29 Jan 2020 20:16 UTC; 20 points) (
- Robustness to fundamental uncertainty in AGI alignment by 3 Mar 2020 23:35 UTC; 14 points) (
- [Article review] Artificial Intelligence, Values, and Alignment by 9 Mar 2020 12:42 UTC; 13 points) (
- RFC: Mental phenomena in AGI alignment by 5 Jul 2018 20:52 UTC; 12 points) (
- Aligned AI May Depend on Moral Facts by 15 Jun 2018 1:33 UTC; 8 points) (
- Robustness to fundamental uncertainty in AGI alignment by 27 Jul 2018 0:41 UTC; 7 points) (
- 31 Jan 2020 0:32 UTC; 3 points) 's comment on Value uncertainty by (
- 1 Feb 2020 5:48 UTC; 3 points) 's comment on Value uncertainty by (
- 16 Feb 2020 1:37 UTC; 3 points) 's comment on Trying for Five Minutes on AI Strategy by (
Why “moral nihilism—the assumption that no moral facts exist—implies the impossibility of building aligned AGI”?
I personally think that alignment is impossible because, shortly speaking, “humans don’t have values”—they have complex unstable behaviour, which could be simplified by using idea of preferences, but there is no actual preferences.
This two claims seems similar: non-existence of values and non-existence of moral facts, but are they actually the same claim?
This is an interesting point to bring up, and maybe one I’ll find a way to explore more.
You’re right that there is a sense in which humans don’t seem to have preferences or values because both are fuzzy categories that draw arbitrary distinctions. The best way to rescue them seems to be to think of values as thoughts that meet at least the following criteria:
orderable (though not in practice totally or even partially or pre orderable)
about experiences (to use your framing, perhaps about possible behaviors)
There might be others but I think that’s enough to let us carve up the space of things that are candidate behaviors such that we exclude the things that are not values and include the things that are. Also this gives a rather odd version of what preferences/values are because it eliminates the normative assumptions normally made when talking about preferences in order to better reflect the “unstable behavior” of humans. I still find value in talking about values, though, because it gives a way to distinguish between the interior experience of value and the exterior experience of acting.
On you’re second point I disagree for the reason just given: there is a sense in which we can talk about values that are non-normative and so avoid the need for moral facts, making them not equivalent. I have, though, neglected to address in this section (or elsewhere in the paper) this take on what we mean by values and made sure it’s clear I’m holding it separate from some other discussions about values/preferences where normativity is sought or assumed.
I have started to write longer text where I will explore the idea of the non-existence of values, and I expect to have ready to present version in 2-3 months, and will share it on LW, so we then could again compare our views on this topic. In it I will adress 3 different ways who we coul learn persons values: 1) his action 2) his emotion 3) his external claims—which could be different from his thoughts about his values, so your idea may be the 4th level.
Great, I look forward to it! That’s a topic I’ve been somewhat unwilling to tackle just now because I’ve identified what I’m working on now as crucial to other lines of thinking I want to explore but consider it crucial to how we’re going to address problems we face in designing alignment schemes.
As you mentionned, no axiology can be inferred from ontology alone.
Even with meta-ethical uncertainty, if we want to build an agent that takes decisions/actions, it needs some initial axiology. If you include (P) “never consider anything as a moral fact” as part of your axology, then two things might happen:
1) This assertion (P) stays in the agent without being modified
2) The agent rewrites its own axiology and modify/delete (P)
I see a problem here. If 1) holds, then it has considered (P) has a moral fact, absurd. If 2) holds, then your agent has lost the meta-ethics principle you wanted him to keep.
So maybe you wanted to put the meta-ethics uncertainty inside the ontology ? It this is what you meant, that doesn’t seem to solve the axiology problem.
It’s not necessarily that we don’t want an agent that never considers something to be a moral fact, but that we are safer if we designed AGI under the assumption that moral facts don’t exist. This does not necessarily mean precluding an AGI coming to believe in the existence of moral facts and almost certainly doesn’t mean giving the AGI a preference to not believe in moral facts, but that alignment schemes should not depend on moral facts existing, and thus should not depend on the ability to make normative assumptions that will be compatible with alignment.