Beyond Astronomical Waste
Faced with the astronomical amount of unclaimed and unused resources in our universe, one’s first reaction is probably wonderment and anticipation, but a second reaction may be disappointment that our universe isn’t even larger or contains even more resources (such as the ability to support 3^^^3 human lifetimes or perhaps to perform an infinite amount of computation). In a previous post I suggested that the potential amount of astronomical waste in our universe seems small enough that a total utilitarian (or the total utilitarianism part of someone’s moral uncertainty) might reason that since one should have made a deal to trade away power/resources/influence in this universe for power/resources/influence in universes with much larger amounts of available resources, it would be rational to behave as if this deal was actually made. But for various reasons a total utilitarian may not buy that argument, in which case another line of thought is to look for things to care about beyond the potential astronomical waste in our universe, in other words to explore possible sources of expected value that may be much greater than what can be gained by just creating worthwhile lives in this universe.
One example of this is the possibility of escaping, or being deliberately uplifted from, a simulation that we’re in, into a much bigger or richer base universe. Or more generally, the possibility of controlling, through our decisions, the outcomes of universes with much greater computational resources than the one we’re apparently in. It seems likely that under an assumption such as Tegmark’s Mathematical Universe Hypothesis, there are many simulations of our universe running all over the multiverse, including in universes that are much richer than ours in computational resources. If such simulations exist, it also seems likely that we can leave some of them, for example through one of these mechanisms:
Exploiting a flaw in the software or hardware of the computer that is running our simulation (including “natural simulations” where a very large universe happens to contain a simulation of ours without anyone intending this).
Exploiting a flaw in the psychology of agents running the simulation.
Altruism (or other moral/axiological considerations) on the part of the simulators.
Other instrumental reasons for the simulators to let out simulated beings, such as wanting someone to talk to or play with. (Paul Christiano’s recent When is unaligned AI morally valuable? contains an example of this, however the idea there only lets us escape to another universe similar to this one.)
(Being run as a simulation in another universe isn’t necessarily the only way to control what happens in that universe. Another possibility is if universes with halting oracles exist (which is implied by Tegmark’s MUH since they exist as mathematical structures in the arithmetical hierarchy), some of their oracle queries may be questions whose answers can be controlled by our decisions, in which case we can control what happens in those universes without being simulated by them (in the sense of being run step by step in a computer). Another example is that superintelligent beings may be able to reason about what our decisions are without having to run a step by step simulation of us, even without access to a halting oracle.)
The general idea here is for a superintelligence descending from us to (after determining that this is an advisable course of action) use some fraction of the resources of this universe to reason about or search (computationally) for much bigger/richer universes that are running us as simulations or can otherwise be controlled by us, and then determine what we need to do to maximize the expected value of the consequences of our actions on the base universes, perhaps through one or more of the above listed mechanisms.
Practical Implications
Realizing this kind of existential hope seems to require a higher level of philosophical sophistication than just preventing astronomical waste in our own universe. Compared to that problem, here we have more questions of a philosophical nature, for which no empirical feedback seems possible. It seems very easy to make a mistake somewhere along the chain of reasoning and waste a more-than-astronomical amount of potential value, for example by failing to realize the possibility of affecting bigger universes through our actions, incorrectly calculating the expected value of such a strategy, failing to solve the distributional/ontological shift problem of how to value strange and unfamiliar processes or entities in other universes, failing to figure out the correct or optimal way to escape into or otherwise influence larger universes, etc.
The total utilitarian in me is thus very concerned about trying to preserve and improve the collective philosophical competence of our civilization, such that when it becomes possible to pursue strategies like ones listed above, we’ll be able to make the right decisions. The best opportunity to do this that I can foresee is the advent of advanced AI, which is another reason I want to push for AIs that are not just value aligned with us, but also have philosophical competence that scales with their other intellectual abilities, so they can help correct the philosophical errors of their human users (instead of merely deferring to them), thereby greatly improving our collective philosophical competence.
Anticipated Questions
How is this idea related to Nick Bostrom’s Simulation Argument? Nick’s argument focuses on the possibility of post-humans (presumably living in a universe similar to ours but just at a later date) simulating us as their ancestors. It does not seem to consider that we may be running as simulations in much larger/richer universes, or that this may be a source of great potential value.
Isn’t this a form of Pascal’s Mugging? I’m not sure. It could be that when we figure out how to solve Pascal’s Mugging it will become clear that we shouldn’t try to leave our simulation for reasons similar to why we shouldn’t pay the mugger. However the analogy doesn’t seem so tight that I think this is highly likely. Also, note that the argument here isn’t that we should do the equivalent of “pay the mugger” but rather that we should try to bring ourselves into a position where we can definitively figure out what the right thing to do is.
- 2018 Review: Voting Results! by 24 Jan 2020 2:00 UTC; 135 points) (
- The Main Sources of AI Risk? by 21 Mar 2019 18:28 UTC; 121 points) (
- Questioning the Foundations of EA by 27 Aug 2022 20:44 UTC; 101 points) (EA Forum;
- Long Reflection Reading List by 24 Mar 2024 16:27 UTC; 92 points) (EA Forum;
- Rationality Exercises Prize of September 2019 ($1,000) by 11 Sep 2019 0:19 UTC; 89 points) (
- Some Thoughts on Metaphilosophy by 10 Feb 2019 0:28 UTC; 76 points) (
- What (standalone) LessWrong posts would you recommend to most EA community members? by 9 Feb 2022 0:31 UTC; 67 points) (EA Forum;
- 20 May 2021 18:23 UTC; 56 points) 's comment on Response to “What does the universal prior actually look like?” by (
- Is there a good place to find the “what we know so far” of the EA movement? by 29 Sep 2019 8:42 UTC; 25 points) (EA Forum;
- 17 Sep 2022 10:38 UTC; 21 points) 's comment on The religion problem in AI alignment by (EA Forum;
- 21 Jun 2019 15:34 UTC; 18 points) 's comment on A case for strategy research: what it is and why we need more of it by (
- 23 May 2023 20:25 UTC; 18 points) 's comment on Worrying less about acausal extortion by (
- [Book Review] “Suffering-focused Ethics” by Magnus Vinding by 28 Dec 2021 5:58 UTC; 15 points) (
- 27 May 2020 9:01 UTC; 14 points) 's comment on AGIs as collectives by (
- 4 Nov 2019 6:03 UTC; 14 points) 's comment on But exactly how complex and fragile? by (
- Alignment Newsletter #46 by 22 Feb 2019 0:10 UTC; 12 points) (
- 21 Aug 2019 1:40 UTC; 11 points) 's comment on What are the reasons to *not* consider reducing AI-Xrisk the highest priority cause? by (
- LW4EA: Beyond Astronomical Waste by 10 May 2022 14:50 UTC; 9 points) (EA Forum;
- 22 Jun 2024 21:31 UTC; 9 points) 's comment on Impartialist Sentientism and Existential Anxiety about Moral Circle Explosion by (EA Forum;
- If I Was An Eccentric Trillionaire by 9 Aug 2023 7:56 UTC; 9 points) (
- 17 Feb 2019 16:34 UTC; 8 points) 's comment on Some disjunctive reasons for urgency on AI risk by (
- 19 Oct 2021 4:23 UTC; 6 points) 's comment on [Book Review] “Suffering-focused Ethics” by Magnus Vinding by (
- 20 Feb 2020 0:14 UTC; 6 points) 's comment on [AN #80]: Why AI risk might be solved without additional intervention from longtermists by (
- 23 Jan 2020 23:01 UTC; 5 points) 's comment on Concerns Surrounding CEV: A case for human friendliness first by (
- 22 Aug 2019 17:40 UTC; 4 points) 's comment on Simulation Argument: Why aren’t ancestor simulations outnumbered by transhumans? by (
- 27 Aug 2019 8:05 UTC; 4 points) 's comment on Six AI Risk/Strategy Ideas by (
- 20 Sep 2019 17:38 UTC; 4 points) 's comment on The strategy-stealing assumption by (
- 17 Sep 2022 1:17 UTC; 3 points) 's comment on Levels of goals and alignment by (
- 15 Oct 2022 7:55 UTC; 2 points) 's comment on Announcing the Future Fund’s AI Worldview Prize by (EA Forum;
- 20 Feb 2020 20:53 UTC; 2 points) 's comment on [AN #80]: Why AI risk might be solved without additional intervention from longtermists by (
- 29 Aug 2019 7:38 UTC; 2 points) 's comment on Six AI Risk/Strategy Ideas by (
- 27 Apr 2024 6:38 UTC; 2 points) 's comment on Eric Neyman’s Shortform by (
- 11 Feb 2019 0:38 UTC; 2 points) 's comment on The Argument from Philosophical Difficulty by (
- 23 Sep 2019 23:28 UTC; 1 point) 's comment on Build a Causal Decision Theorist by (
I think that at the time this post came out, I didn’t have the mental scaffolding necessary to really engage with it – I thought of this question as maybe important, but sort of “above my paygrade”, something better left to other people who would have the resources to engage more seriously with it.
But, over the past couple years, the concepts here have formed an important component of my understanding of robust agency. Much of this came from private in-person conversations, but this post is the best writeup of the concept I’m currently aware of.
One thing I like about this post is the focus on philosophical competence. Previously, I’d thought of this question as dangerous to think about, because you might make philosophical mistakes that doomed you or your universe for (in retrospect) silly reasons.
My current model is more like “no, you-with-your-21st-century-human-brain shouldn’t actually attempt to take actions aiming primarily to affect other universes on the macro scale. Negotiating with other universes is something you do when you’re a literal galaxy brain that is quite confident in it’s philosophy.”
But, meanwhile, it seems that:
(note: low-to-mid confidence, still working through these problems myself, and I very much still philosophically confused about at least some of this)
– becoming philosophically competent, as a species, may be one of the most important goals facing humanity, and how this project interfaces with AI development may be crucially important. This may be relevant to people (like me) who aren’t directly working on alignment but trying to have a good model of the strategic landscape.
– a concept not from this particular post, but relevant, is the notion that the question is not “are you in a simulation or not?”, it’s more like “to what degree are you in simulations? Which distribution of agents are you making choices on behalf of?”. And this has some implications about how you should make choices that you should worry about now, before you are a literal galaxy brain. (many of which are more mundane)
– I think there may be a connection between “Beyond Astronominal Waste” and “Robustness to scale.” You can’t be a galaxy brain now, but you can be the sort of person who would be demonstrably safe to scale up, in a way that simulations can detect, which might let you punch above your current weight, in terms of philosophical competence.
This is a post that’s stayed with me since it was published. The title is especially helpful as a handle. It is a simple reference for this idea, that there are deeply confusing philosophical problems that are central to our ability to attain most of the value we care about (and that this might be a central concern when thinking about AI).
It’s not been very close to areas I think about a lot, so I’ve not tried to build on it much, and would be interested in a review from someone who thinks in more detail about these matters more, but I expect they’ll agree it’s a very helpful post to exist.