This argument, as (perhaps incompletely) stated, also works for predictive processing; reductio ad absurdum?
I think predictive processing has the same problem as reward if you are part of the updated model rather than the model being a modular part of you. It’s a change to your own self that’s not your decision (not something endorsed), leading to value drift and other undesirable deterioration. So for humans, it’s a real problem, just not the most urgent one. Of course, there is no currently feasible alternative, but neither is there an alternative for reward in RL.
I think predictive processing has the same problem as reward if you are part of the updated model rather than the model being a modular part of you. It’s a change to your own self that’s not your decision (not something endorsed), leading to value drift and other undesirable deterioration. So for humans, it’s a real problem, just not the most urgent one. Of course, there is no currently feasible alternative, but neither is there an alternative for reward in RL.