Personally, my understanding is based on what might be a fundamentally different theory of mind. I believe there’s two major optimization algorithms at work.
Optimizer 1 is a real-time world model prediction error minimizer. Think predictive coding.
That’s my theory of mind. You describe two competing reward systems. But reward systems belong in the domain of Optimizer 2. The way I look at things, meditation (temporarily?) shuts down Optimizer 2, which allows Optimizer 1 to self-optimize unimpeded.
Personally, my understanding is based on what might be a fundamentally different theory of mind. I believe there’s two major optimization algorithms at work.
Optimizer 1 is a real-time world model prediction error minimizer. Think predictive coding.
Optimizer 2 is is a operant reinforcement reward system. Optimizer 2 is parasitic on Optimizer 1. The conflict between Optimizer 1 and Optimizer 2 is a mathematical constraint inherent to embedded world optimizers.
That’s my theory of mind. You describe two competing reward systems. But reward systems belong in the domain of Optimizer 2. The way I look at things, meditation (temporarily?) shuts down Optimizer 2, which allows Optimizer 1 to self-optimize unimpeded.