Sure, that’s a sane Oracle. The Weird Oracle used in so many thought experiments doesn’t say “”The taxi will arrive in one minute!”, it says “You will grab your coat in time for the taxi.”.
No, this is an important point: the agent normally doesn’t know the correctness scope of the Oracle’s prediction. It’s only guaranteed to be correct on the actual decision, and can be incorrect in all other counterfactuals. So if the agent knows the boundaries of the correctness scope, they may play chicken and render the Oracle wrong by enacting the counterfactual where the prediction is false. And if the agent doesn’t know the boundaries of the prediction’s correctness, how are they to make use of it in evaluating counterfactuals?
It seems that the way to reason about this is to stipulate correctness of the prediction in all counterfactuals, even though it’s not necessarily correct in all counterfactuals, in the same way as the agent’s decision that is being considered is stipulated to be different in different counterfactuals, even though the algorithm forces it to be the same. So it’s a good generalization of the problem of formulating counterfactuals, it moves the intervention point from agent’s own decisions to correctness of powerful predictors’ claims. These claims act on the counterfactuals generated by the agent’s own decisions, not on the counterfactuals generated by delivery of possible claims, so it’s not about merely treating predictors as agents, it’s a novel setup.
Is there an ELI5 doc about what’s “normal” for Oracles, and why they’re constrained in that way? The examples I see confuse me in that they are exploring what seem like edge cases, and I’m missing the underlying model that makes these cases critical.
Specifically, when you say “It’s only guaranteed to be correct on the actual decision”, why does the agent not know what “correct” means for the decision?
Specifically, when you say “It’s only guaranteed to be correct on the actual decision”, why does the agent not know what “correct” means for the decision?
The agent knows what “correct” means, correctness of a claim is defined for the possible worlds that the agent is considering while making its decision (which by local tradition we confusingly collectively call “counterfactuals”, even though one of them is generated by the actual decision and isn’t contrary to any fact).
In the post Chris_Leong draws attention to the point that since the Oracle knows which possible world is actual, there is nothing forcing its prediction to be correct on the other possible worlds that the agent foolishly considers, not knowing that they are contrary to fact. And my point in this thread is that despite the uncertainty it seems like we have to magically stipulate correctness of the Oracle on all possible worlds in the same way that we already magically stipulate the possibility of making different decisions in different possible worlds, and this analogy might cast some light on the nature of this magic.
That’s an interesting point. I suppose it might be viable to acknowledge that the problem taken literally doesn’t require the prediction to be correct outside of the factual, but nonetheless claim that we should resolve the vagueness inherent in the question about what exactly the counterfactual is by constructing it to meet this condition. I wouldn’t necessarily be strongly against this—my issue is confusion about what an Oracle’s prediction necessarily entails.
No, I suspect it’s a correct ingredient of counterfactuals, one I didn’t see discussed before, not an error restricted to a particular decision theory. There is no contradiction in considering each of the counterfactuals as having a given possible decision made by the agent and satisfying the Oracle’s prediction, as the agent doesn’t know that it won’t make this exact decision. And if it does make this exact decision, the prediction is going to be correct, just like the possible decision indexing the counterfactual is going to be the decision actually taken. Most decision theories allow explicitly considering different possible decisions, and adding correctness of the Oracle’s prediction into the mix doesn’t seem fundamentally different in any way, it’s similarly sketchy.
Thanks for patience with this. I am still missing some fundamental assumption or framing about why this is non-obvious (IMO, either the Oracle is wrong, or the choice is illusory). I’ll continue to examine the discussions and examples in hopes that it will click.
I presume Vladimir and me are likely discussing this from within the determinist paradigm in which “either the Oracle is wrong, or the choice is illusory” doesn’t apply (although I propose a similar idea in Why 1-boxing doesn’t imply backwards causation).
IMO, either the Oracle is wrong, or the choice is illusory
This is similar to determinism vs. free will, and suggests the following example. The Oracle proclaims: “The world will follow the laws of physics!”. But in the counterfactual where an agent takes a decision that won’t actually be taken, the fact of taking that counterfactual decision contradicts the agent’s cognition following the laws of physics. Yet we want to think about the world within the counterfactual as if the laws of physics are followed.
Sure, that’s a sane Oracle. The Weird Oracle used in so many thought experiments doesn’t say “”The taxi will arrive in one minute!”, it says “You will grab your coat in time for the taxi.”.
No, this is an important point: the agent normally doesn’t know the correctness scope of the Oracle’s prediction. It’s only guaranteed to be correct on the actual decision, and can be incorrect in all other counterfactuals. So if the agent knows the boundaries of the correctness scope, they may play chicken and render the Oracle wrong by enacting the counterfactual where the prediction is false. And if the agent doesn’t know the boundaries of the prediction’s correctness, how are they to make use of it in evaluating counterfactuals?
It seems that the way to reason about this is to stipulate correctness of the prediction in all counterfactuals, even though it’s not necessarily correct in all counterfactuals, in the same way as the agent’s decision that is being considered is stipulated to be different in different counterfactuals, even though the algorithm forces it to be the same. So it’s a good generalization of the problem of formulating counterfactuals, it moves the intervention point from agent’s own decisions to correctness of powerful predictors’ claims. These claims act on the counterfactuals generated by the agent’s own decisions, not on the counterfactuals generated by delivery of possible claims, so it’s not about merely treating predictors as agents, it’s a novel setup.
Is there an ELI5 doc about what’s “normal” for Oracles, and why they’re constrained in that way? The examples I see confuse me in that they are exploring what seem like edge cases, and I’m missing the underlying model that makes these cases critical.
Specifically, when you say “It’s only guaranteed to be correct on the actual decision”, why does the agent not know what “correct” means for the decision?
The agent knows what “correct” means, correctness of a claim is defined for the possible worlds that the agent is considering while making its decision (which by local tradition we confusingly collectively call “counterfactuals”, even though one of them is generated by the actual decision and isn’t contrary to any fact).
In the post Chris_Leong draws attention to the point that since the Oracle knows which possible world is actual, there is nothing forcing its prediction to be correct on the other possible worlds that the agent foolishly considers, not knowing that they are contrary to fact. And my point in this thread is that despite the uncertainty it seems like we have to magically stipulate correctness of the Oracle on all possible worlds in the same way that we already magically stipulate the possibility of making different decisions in different possible worlds, and this analogy might cast some light on the nature of this magic.
That’s an interesting point. I suppose it might be viable to acknowledge that the problem taken literally doesn’t require the prediction to be correct outside of the factual, but nonetheless claim that we should resolve the vagueness inherent in the question about what exactly the counterfactual is by constructing it to meet this condition. I wouldn’t necessarily be strongly against this—my issue is confusion about what an Oracle’s prediction necessarily entails.
Regarding, your notion about things being magically stipulated, I suppose there’s some possible resemblance there with the ideas I proposed before in Counterfactuals As A Matter of Social Convention, although The Nature of Counterfactuals describes where my views have shifted to since then.
Hmm. So does this only apply to CDT agents, who foolishly believe that their decision is not subject to predictions?
No, I suspect it’s a correct ingredient of counterfactuals, one I didn’t see discussed before, not an error restricted to a particular decision theory. There is no contradiction in considering each of the counterfactuals as having a given possible decision made by the agent and satisfying the Oracle’s prediction, as the agent doesn’t know that it won’t make this exact decision. And if it does make this exact decision, the prediction is going to be correct, just like the possible decision indexing the counterfactual is going to be the decision actually taken. Most decision theories allow explicitly considering different possible decisions, and adding correctness of the Oracle’s prediction into the mix doesn’t seem fundamentally different in any way, it’s similarly sketchy.
Thanks for patience with this. I am still missing some fundamental assumption or framing about why this is non-obvious (IMO, either the Oracle is wrong, or the choice is illusory). I’ll continue to examine the discussions and examples in hopes that it will click.
I presume Vladimir and me are likely discussing this from within the determinist paradigm in which “either the Oracle is wrong, or the choice is illusory” doesn’t apply (although I propose a similar idea in Why 1-boxing doesn’t imply backwards causation).
This is similar to determinism vs. free will, and suggests the following example. The Oracle proclaims: “The world will follow the laws of physics!”. But in the counterfactual where an agent takes a decision that won’t actually be taken, the fact of taking that counterfactual decision contradicts the agent’s cognition following the laws of physics. Yet we want to think about the world within the counterfactual as if the laws of physics are followed.