In game theory, there are a number of situations where it is rational to handicap your own rationality: Reduce your number of choices, take away information, etc.
TDT is intended to eliminate this. A TDT-agent—one that’s correctly modeled by the environment, not that some other agent thinks is a CDT-agent—is supposed to never benefit from having any option taken away from it, and will never pay to avoid learning a piece of information.
Er, this is assuming that the information revealed is not intentionally misleading, correct? Because certainly you could give a TDT agent an extra option which would be rational to take on the basis of the information available to the agent, but which would still be rigged to be worse than all other options.
Or in other words, the TDT agent can never be aware of such a situation.
TDT is intended to eliminate this. A TDT-agent—one that’s correctly modeled by the environment, not that some other agent thinks is a CDT-agent—is supposed to never benefit from having any option taken away from it, and will never pay to avoid learning a piece of information.
Er, this is assuming that the information revealed is not intentionally misleading, correct? Because certainly you could give a TDT agent an extra option which would be rational to take on the basis of the information available to the agent, but which would still be rigged to be worse than all other options.
Or in other words, the TDT agent can never be aware of such a situation.
Amendment accepted.