What if you believe in DKRUDT, the “Dick Kick’em rewards you” decision theory?
Seriously though, Newcomb’s setup is not adversarial in the same way, the predictor rewards or punishes you for actions, not beliefs. Your internal reasoning does not matter, as long as you end up one-boxing you walk away with more money.
Seriously though, Newcomb’s setup is not adversarial in the same way, the predictor rewards or punishes you for actions, not beliefs.
This cannot be true becuase it would violate cause and effect. The predictor will decide to reward/punish you with the amount of money put in the boxes. This reward/punishment is done BEFORE any action is taken, and so it is based purely on the beliefs of the agent. If it were based on the actions of the agent, that would mean that the cause of the reward/punishment happens AFTER the decision was made, which violates cause and effect. The cause must come BEFORE.
Let me clarify what I said. Any decision theory or no decision theory at all that results in someone one-boxing is rewarded. Examples: Someone hates touching transparent boxes. Someone likes a mystery of an opaque box. Someone thinking that they don’t deserve a guaranteed payout and hoping for an empty box. Someone who is a gambler. Etc. What matters is the outcome, not the thought process.
That just means the AI cares about a particular class of decision theories rather than a specific one like Dick Kick’em. I could re-run the same thought experiment but instead Dick Kick’em says:
”I am going to read you mind and if you believe in a decision theory that one-boxes in Newcomb’s Paradox I will leave you alone, but if you believe in any other decision theory I will kick you in the dick”
In this variation, Dick Kick’em would be judging the agent based on the exact same criterea that the AI in Newcomb’s problem is using. All I have done is remove the game afterwards but that is somewhat irrelevant since the AI doesn’t judge you on your actions, just what you would do if you were in a Newcomb-type scenario.
“I am going to read you mind and if you believe in a decision theory that one-boxes in Newcomb’s Paradox I will leave you alone, but if you believe in any other decision theory I will kick you in the dick”
Sure, that’s possible. Assuming there are no Newcomb’s predictors in that universe, but only DK, rational agents believe in two-boxing. I am lost as to how it is related to your original point.
What if you believe in DKRUDT, the “Dick Kick’em rewards you” decision theory?
Seriously though, Newcomb’s setup is not adversarial in the same way, the predictor rewards or punishes you for actions, not beliefs. Your internal reasoning does not matter, as long as you end up one-boxing you walk away with more money.
This cannot be true becuase it would violate cause and effect. The predictor will decide to reward/punish you with the amount of money put in the boxes. This reward/punishment is done BEFORE any action is taken, and so it is based purely on the beliefs of the agent. If it were based on the actions of the agent, that would mean that the cause of the reward/punishment happens AFTER the decision was made, which violates cause and effect. The cause must come BEFORE.
Let me clarify what I said. Any decision theory or no decision theory at all that results in someone one-boxing is rewarded. Examples: Someone hates touching transparent boxes. Someone likes a mystery of an opaque box. Someone thinking that they don’t deserve a guaranteed payout and hoping for an empty box. Someone who is a gambler. Etc. What matters is the outcome, not the thought process.
That just means the AI cares about a particular class of decision theories rather than a specific one like Dick Kick’em. I could re-run the same thought experiment but instead Dick Kick’em says:
”I am going to read you mind and if you believe in a decision theory that one-boxes in Newcomb’s Paradox I will leave you alone, but if you believe in any other decision theory I will kick you in the dick”
In this variation, Dick Kick’em would be judging the agent based on the exact same criterea that the AI in Newcomb’s problem is using. All I have done is remove the game afterwards but that is somewhat irrelevant since the AI doesn’t judge you on your actions, just what you would do if you were in a Newcomb-type scenario.
Sure, that’s possible. Assuming there are no Newcomb’s predictors in that universe, but only DK, rational agents believe in two-boxing. I am lost as to how it is related to your original point.