OOOOOOH! I think I see (part of, but not all) of the misunderstanding here. I wasn’t talking about how Omega can take this into account, I was talking about how the agent Omega is playing games with would take this into account.
ie, not how Omega deals with the problem, but how I would.
Problems involving Omega probably aren’t useful examples for demonstrating your problem either way since Omega will accurately predict our actions either way and our identity angst is irrelevant.
I’d like to see an instantiation of the type of problem you mentioned above, involving the many explicitly dependant systems. Something involving a box to pick or a bet to take. Right now the requirements of the model are not defined much beyond ‘apply standard decision theory with included mechanism for handling uncertainty at such time as the problem becomes available’.
OOOOOOH! I think I see (part of, but not all) of the misunderstanding here. I wasn’t talking about how Omega can take this into account, I was talking about how the agent Omega is playing games with would take this into account.
ie, not how Omega deals with the problem, but how I would.
Problems involving Omega probably aren’t useful examples for demonstrating your problem either way since Omega will accurately predict our actions either way and our identity angst is irrelevant.
I’d like to see an instantiation of the type of problem you mentioned above, involving the many explicitly dependant systems. Something involving a box to pick or a bet to take. Right now the requirements of the model are not defined much beyond ‘apply standard decision theory with included mechanism for handling uncertainty at such time as the problem becomes available’.
So? The graph still handles that.