Why does this approach only need to be implemented in neo-cortex like AGIs? If we have a factored series of value functions in an RL agent then we should be able to take the same approach? But I guess you are thinking that the basal ganglia learning algorithms already do this for us so it is a convenient approach?
Side note. I found the distinction between confusion and conflict a bit… confusing! Confusion here is the agent updating a belief while conflict is the agent deciding to take an action?
Why does this approach only need to be implemented in neo-cortex like AGIs?
Oh, I wasn’t saying that. I wouldn’t know either way, I haven’t thought about it. RL is a very big and heterogeneous field. I only know little bits and pieces of it. It’s a lot easier to make a specific proposal that applies to a specific architecture—the architecture that I happen to be familiar with—than to try to make a more general proposal. So that’s all I did.
What do you mean by “factored series of value functions”? If you’re thinking of my other post, maybe that’s possible, although not what I had in mind, because humans can feel conflicted, but my other post is talking about a mechanism that does not exist in the human brain.
Confusion here is the agent updating a belief while conflict is the agent deciding to take an action?
Why does this approach only need to be implemented in neo-cortex like AGIs? If we have a factored series of value functions in an RL agent then we should be able to take the same approach? But I guess you are thinking that the basal ganglia learning algorithms already do this for us so it is a convenient approach?
Side note. I found the distinction between confusion and conflict a bit… confusing! Confusion here is the agent updating a belief while conflict is the agent deciding to take an action?
Oh, I wasn’t saying that. I wouldn’t know either way, I haven’t thought about it. RL is a very big and heterogeneous field. I only know little bits and pieces of it. It’s a lot easier to make a specific proposal that applies to a specific architecture—the architecture that I happen to be familiar with—than to try to make a more general proposal. So that’s all I did.
What do you mean by “factored series of value functions”? If you’re thinking of my other post, maybe that’s possible, although not what I had in mind, because humans can feel conflicted, but my other post is talking about a mechanism that does not exist in the human brain.
Yeah, that’s what I was going for.