Disclaimer: this question discusses potential infohazards, which may cause harm or psychological distress if true. If you decide to read this post, please be warned and do so at your own discretion.
Hi, I am someone new to LessWrong so please excuse any mistakes I may make in my discussions (you’re more than welcome to point them out). This question is intended to ask whether or not acausal extortion is possible and to pose a scenario which I am unsure about, hence please feel free to critique my points and arguments.
Acausal trade is essentially when two agents simulate each other to predict each other’s actions, and acausal extortion (or blackmail) is when one agent applies a negative incentive to influence the behavior of another agent when the latter agent models the former agent. From what I’ve gathered, acausal trade (and by extension acausal extortion) between humans and artificial superintelligences (ASIs) is impossible as it requires lots of computing power on the human’s end in order for the ASI to not have an incentive to defect instead cooperate. However, this still leaves open the possibility of a non-superintelligence (such as an alien) who has control over an ASI using it to acausally extort us. To see what I mean, consider the scenario below:
On some branch of the multiverse, there exists an alien with control over an aligned ASI. The alien then asks the ASI to create many simulations of the other branches of the multiverse and you are contained in one of these simulations. The alien observes you and realizes that you realize the possibility of such a scenario taking place (since you are reading this paragraph). The alien then thinks: If you do not take action X, it will ask the ASI to create an arbitrarily large number of simulations of you and torture them for an arbitrarily long period of time. And since these simulations have the exact same subjective experience as you, you are compelled to carry out action X as you are uncertain about whether or not you’re in a simulation (indexical uncertainty). And since it is much easier to model non-superintelligences than ASIs, this scenario avoids the pitfall of scenarios like Roko’s Basilisk where humans are limited by our computing power.
The scenario above is partly inspired by the famous Roko’s Basilisk and Stuart Armstrong’s AI in a Box thought experiment.
Now I’d like to address some of the common refutations to acausal extortion scenarios and why I don’t think any of them work very well against this particular scenario. (Though it’s very likely I have made mistakes or engaged strawman versions of these arguments, and if so, please do not hesitate to point them out).
Yeah, sure, this scenario could happen, but it’s very very very unlikely right?
The above scenario, however unlikely, is guaranteed to take place with probability 1 if the Many-Worlds Interpretation of quantum mechanics is correct, as every possible outcome occurs in such a multiverse. And even if the measure of this particular branch of the multiverse is extremely low, if the alien creates an extremely large number of simulations to create indexical uncertainty, it can still have a significant effect on your most probable environment and possible future observer-moments.
Ok, sure, but I can imagine any action X and there will be a quantum branch where the alien wants you to take that particular action, this is just a Pascal’s Mugging where the many-gods refutation still applies.
The action X can be indeed be any arbitrary action, but such aliens are likely to have convergent instrumental goals such that the probability distribution of X will not be random. And I don’t find the many-gods refutation to be satisfactory since, well, it just concludes that you will be tortured no matter what you do (not a good ending).
Hmm, well then if I commit to ignore any such attempts of blackmail, then the blackmailer will have no incentive to employ such a tactic.
Sure, it works in principle, but humans cannot commit to an action with 100% certainty. How confident are you that none of your bazillions of copies will give in? And also, in the MWI every possible outcome occurs, including outcomes where you do give in to the blackmailer’s demand, so it’s likely that doing so will yield positive expected utility (however small) for the blackmailer.
In the above section I addressed some of the refutations that I hear commonly but do not find all that convincing. I am not sure how much I believe in this whole thing yet but since I am not an expert in decision theory and this sort of thing in general I wanted to hear some criticism from more knowledgeable people on the scenario I’ve posed as well as on how to avoid feeling anxious over acausal extortion. Thanks in advance.
It is true that superinteligences are much more difficult to model, and therefore predict and give them a reason to actually carry out threats is low, but its not like its easy all of a sudden to say that “ah, aliens or humans arent as complex, therefore i should take them seriously.” Its virtually impossible to know with any real accuracy what aliens are going to do, because we know nothing about them, and even humans, are diverse and complicated in their own right, so good luck trying to predict or even pretend to know what they are going to think or do without any real evidance.
I think I basically agree, though I am pretty uncertain. You’d basically have to simulate not just the other being, but also the other being simulating you, with a certain fidelity. In my post I posed the scenario where the other being is watching you through an ASI simulation, and so it is much more computationally easier for them to simulate you in their head, but this means you have to simulate what the other being is thinking as well as what it is seeing. Simply modelling the being as thinking “I will torture him for X years if he doesn’t do action Y” is an oversimplification since you also have to expand out the “him” as “a simulation of you” in very high detail.
Therefore, I think it is still extremely computation-intensive for us to simulate the being simulating us.
This reply i think is correct, but let me add something: the number of possible ways you can simulate a being is immense; and wich one should you choose? You have no idea and no evidence to judge from. And even when you are taking an action, how would you know if its the right one? So why even engage, there are also many other entities that might exist that dont even do acausal trade and might still do something to you. It seems to be the case that the best way of action is to just ignore these cases, cuz if you forget about them they will conclude that youre impenetrable, and so no reason to follow up on anything.
Tldr: just cuz its “easyer”, doesnt mean its real.
Hi, ive recently stumbled upon this post and am a bit worried. Should i be?
Hey, I wouldn’t worry about it. I don’t think anything productive will come of that.
Can you go in more detail?
Maybe we’ll write up an FAQ on the topic, not sure, but I still wouldn’t worry.
Yes please, I think that would be quite helpful. I’m no longer that scared of it but still has some background anxiety sometimes flaring up. I feel like an FAQ or at least some form of “official” explanation from knowledgeable ppl of why it’s not a big deal would help a lot. :)
Comepletely agree.