Although you can’t take it because of the pre-commitment mechanism.
This is a crux for me. In such worlds where this prediction is possible, you can no longer say “because of” and really know that’s true. I suspect the precommittment mechanism is the way you KNOW that you can’t take the box, but it’s not why you can’t take the box.
I don’t really get that. For example, you could put a cryptographic lock on the box (let’s assume there is no way around it without the key), and then throw away the key. It seems that now you actually are not able to access the box, because you do not have the key. And you can also at the same time know that this is the case.
Sure, there are any number of commitment mechanisms which would be hard (or NP-hard) to bypass. If the prediction and box-content selection was performed by Omega based on that cause, then fine. If instead, it was based on a more complete modeling of the universe, REGARDLESS of whether the visible mechanism “could” be bypassed, then there are other causes than that mechanism.
There could be but there does not need to be, I would say. Or maybe I really do not get what you are talking about. It could really be that if the cryptographic lock was not in place, that then you could take the box, and there is nothing else that prevents you from doing this. I guess I have an implicit model where I look at the world from a cartesian perspective. So is what you’re saying about counterfactuals, and that I am using them in a way that is not valid, and that I do not acknowledge this?
I think my main point is that “because” is a tricky word to use normally, and gets downright weird in a universe that includes Omega levels of predictions about actions that feel “free” from the agent.
If Omega made the prediction, that means Omega sees the actual future, regardless of causality or intent or agent-visible commitment mechanisms.
This is a crux for me. In such worlds where this prediction is possible, you can no longer say “because of” and really know that’s true. I suspect the precommittment mechanism is the way you KNOW that you can’t take the box, but it’s not why you can’t take the box.
I don’t really get that. For example, you could put a cryptographic lock on the box (let’s assume there is no way around it without the key), and then throw away the key. It seems that now you actually are not able to access the box, because you do not have the key. And you can also at the same time know that this is the case.
Not sure why this should be impossible to say.
Sure, there are any number of commitment mechanisms which would be hard (or NP-hard) to bypass. If the prediction and box-content selection was performed by Omega based on that cause, then fine. If instead, it was based on a more complete modeling of the universe, REGARDLESS of whether the visible mechanism “could” be bypassed, then there are other causes than that mechanism.
There could be but there does not need to be, I would say. Or maybe I really do not get what you are talking about. It could really be that if the cryptographic lock was not in place, that then you could take the box, and there is nothing else that prevents you from doing this. I guess I have an implicit model where I look at the world from a cartesian perspective. So is what you’re saying about counterfactuals, and that I am using them in a way that is not valid, and that I do not acknowledge this?
I think my main point is that “because” is a tricky word to use normally, and gets downright weird in a universe that includes Omega levels of predictions about actions that feel “free” from the agent.
If Omega made the prediction, that means Omega sees the actual future, regardless of causality or intent or agent-visible commitment mechanisms.