I agree that what you’re describing is a valid way of looking at what’s going on—it’s just not the way I think about it, since I find that it’s not very helpful to think of a model as a subagent of gradient descent, as gradient descent really isn’t itself an agent in a meaningful sense, nor do I think it can really be understood as “trying” to do anything in particular.
Sure, makes sense! Though to be clear, I believe what I’m describing should apply to optimizers other than just gradient descent — including optimizers one might think of as reward-maximizing agents.
I agree that what you’re describing is a valid way of looking at what’s going on—it’s just not the way I think about it, since I find that it’s not very helpful to think of a model as a subagent of gradient descent, as gradient descent really isn’t itself an agent in a meaningful sense, nor do I think it can really be understood as “trying” to do anything in particular.
Sure, makes sense! Though to be clear, I believe what I’m describing should apply to optimizers other than just gradient descent — including optimizers one might think of as reward-maximizing agents.