In my motivations and in my decision theory, dynamic inconsistency is Always Wrong. Among other things, it always implies an agent unstable under reflection.
If you really want to impress an inspector who can see your internal state, by altering your utility function to conform to their wishes, then one strategy would be to create a trusted external “brain surgeon” agent with the keys to your utility function to change it back again after your utility function has been inspected—and then forget all about the existence of the surgeon.
The inspector will be able to see the lock on your utility function—but those are pretty standard issue.
If you really want to impress an inspector who can see your internal state, by altering your utility function to conform to their wishes, then one strategy would be to create a trusted external “brain surgeon” agent with the keys to your utility function to change it back again after your utility function has been inspected—and then forget all about the existence of the surgeon.
The inspector will be able to see the lock on your utility function—but those are pretty standard issue.