FWIW—here (finally) is the related post I mentioned, which motivated this observation: Natural Abstraction: Convergent Preferences Over Information Structures The context is a power-seeking-style analysis of the naturality of abstractions, where I was determined to have transitive preferences.
It had quite a bit of scope creep already, so I ended up not including a general treatment of the (transitive) ‘sum over orbits’ version of retargetability (and some parts I considered only optimality—sorry! still think it makes sense to start there first and then generalize in this case). The full translation also isn’t necessarily as easy as I thought—it turns out that is transitive specifically for binary functions, so the other cases may not translate as easily as . After noticing that I decided to leave the general case for later.
I did use the sum-over-orbits form, though; which turns out to describe the preferences shared by every “-invariant” distribution over utility functions. Reading between the lines shows roughly what it would look like.
I also moved from to any - not sure if you looked at that, but at least the parts I was using all seem to work just as well with any subgroup. This gives preferences shared by a larger set of distributions, e.g. for an MDP you could in some cases have preferred to for all priors on that are merely invariant to permuting and (rather than requiring them to be invariant to all permutations of utilities).
A while back I was looking for toy examples of environments with different amounts of ‘naturalness’ to their abstractions, and along the way noticed a connection between this version of Gooder Regulator and the Blackwell order.
Inspired by this, I expanded on this perspective of preferences-over-models / abstraction a fair bit here.
It includes among other things:
the full preorder of preferences-shared-by-all-agents over maps X→M(vs. just the maximum)
an argument that actually we want to generalize to this diagram instead[1] :
an extension to preferences shared by ‘most’ agents based on Turner’s power-seeking theorems (and a formal connection to power-seeking)
Personally I think these results are pretty neat; I hope they might be of interest.
I also make one other argument there that directly responds to this post, re: the claim:
I think this sort of true, depending how you interpret it, but really it’s more accurate to say the whole regulator X→M→R has to encode the posterior, not just X→M.
Specifically, as you show, M(X) and (s↦P[S=s|X]) must induce the same equivalence classes of x. This means that they are isomorphic as functions in a particular sense.
But it turns out: lots of (meaningfully) different systems X→S can lead to the same optimal model M(x). This means you can’t recover the full system transition function just by looking at M(x); so it doesn’t really store the whole posterior.
On the other hand, you can uniquely recover the whole posterior (up to technicalities) from M(x) plus the agent’s optimal strategies π:(M,Y)→ΔR. So it is still fair to say the agent as a whole most model the Bayesian posterior; but I’d say it’s not just the model M(X) which does it.
Which in this perspective basically means allowing the ‘system’ or ‘world’ we study to include restrictions on the outcome mapping, e.g. that (s,a), and (s′,a′) lead to the same outcome, which must be given a common utility by any given game. Looking back at this again, I’m not sure I described the distinction quite right in my post (since this setting as you gave it already has a Z distinct from u(Z)), but there is still a distinction.