In much the same way, estimates of value and calculations based on the number of permutations of atoms shouldn’t be mixed together. There being a googleplex possible states in no way implies that any of them have a value over 3 (or any other number). It does not, by itself, imply that any particular state is better than any other. Let alone that any particular state should have value proportional to the total number of states possible.
Restricting yourself to atoms within 8000 light years, instead of the galaxy, just compounds the problem as well, but you noted that yourself. The size of the galaxy wasn’t actually a relevant number, just a (maybe) useful comparison. It’s like when people say that chess has more possible board states than there are atoms in the observable universe times the number of seconds since the Big Bang. It’s not that there’s any specifically useful interaction between atoms and seconds and chess, it’s just to recognize the scale of the problem.
In a Newcombless problem, where you can either have $1,000 or refuse it and have $1,000,000, you could argue that the rational choice is to take the $1,000,000, and then go back for the $1,000 when people’s backs were turned, but it would seem to go against the nature of the problem.
In much the same way, if Omega is a perfect predictor, there is no possible world where you receive $1,000,000 and still end up going back for the second. Either Rachel wouldn’t have objected, or the argument would’ve taken more than 5 minutes, and the boxes disappear, or something.
I’m not sure how Omega factors in the boxes’ contents in this “delayed decision” version. Like, let’s say Irene is will, absent external forces, one box, and Rachel, if Irene receives $1,000,000, will threaten Irene sufficiently to take the second box, and will do nothing if Irene receives nothing. (Also they’re automatons, and these are descriptions of their source code, and so no other unstated factors are able to be taken into account)
Omega simulates reality A, with the box full, sees that Irene will 2 box after threat by Rachel.
Omega simulates reality B, with the box empty, and sees that Irene will 1 box.
Omega, the perfect predictor, cannot make a consistent prediction, and, like the unstoppable force meeting the immovable object, vanishes in a puff of logic.
I think, if you want to aim at this sort of thing, the better formulation is to just claim that Omega is 90% accurate. Then there’s no (immediate) logical contradiction in receiving the $1,000,000 and going back for the second box. And the payoffs should still be correct.
1 box: .9*1,000,000 + .1*0 = 900,000
2 box: .9*1,000 + .1*1,001,000 = 101,000
I expect that this formulation runs folly of what was discussed in this post around the Smoking Lesions problem, where repeated trials may let you change things you shouldn’t be able to (in their example, if you chose to smoke every time, then if the correlation between smoking and lesions was held, then you can change the base rate of the lesions).
That is, I expect that if you ran repeated simulations, to try things out, then strategies like “I will one box, and iff it is full, then I will go back for the second box” will make it so Omega is incapable of predicting at the proposed 90% rate.
I think all of these things might be related to the problem of embedded agency, and people being confused (even if they don’t put it in these terms) that they have an atomic free will that can think about things without affecting or being affected by the world. I’m having trouble resolving this confusion myself, because I can’t figure out what Omega’s prediction looks like instead of vanishing in a puff of logic. It may just be that statements like “I will turn the lever on if, and only if, I expect the lever to be off at the end” are a nonsense decision criteria. But the problem as stated doesn’t seem like it should be impossible, so… I am confused.