About the “right hand rule” agent, I feel it depends on whether it is a hard-coded agent or a learning agent. If it is hard-coded, then clearly it doesn’t require a model. But if it learns such a rule, I would assume it was inferred from a learned model of what mazes are.
For the non-adaptative agent, you say it is less goal-directed; do you see goal-directedness as a continuous spectrum, as a set of zones on this spectrum, or as a binary threshold on this spectrum?
About the “right hand rule” agent, I feel it depends on whether it is a hard-coded agent or a learning agent.
Yes, I meant the hard-coded one. It still seems somewhat goal-directed to me.
do you see goal-directedness as a continuous spectrum, as a set of zones on this spectrum, or as a binary threshold on this spectrum?
Oh, definitely a continuous spectrum. (Though I think several people disagree with me on this, and see it more like a binary-ish threshold. Such people often say things like “intelligence and generalization require some sort of search-like cognition”. I don’t understand their views very well.)
Uh, not really. The mesa optimizers sequence sort of comes from this viewpoint, as does this question, but I haven’t really seen any posts arguing for this position.
About the “right hand rule” agent, I feel it depends on whether it is a hard-coded agent or a learning agent. If it is hard-coded, then clearly it doesn’t require a model. But if it learns such a rule, I would assume it was inferred from a learned model of what mazes are.
For the non-adaptative agent, you say it is less goal-directed; do you see goal-directedness as a continuous spectrum, as a set of zones on this spectrum, or as a binary threshold on this spectrum?
Yes, I meant the hard-coded one. It still seems somewhat goal-directed to me.
Oh, definitely a continuous spectrum. (Though I think several people disagree with me on this, and see it more like a binary-ish threshold. Such people often say things like “intelligence and generalization require some sort of search-like cognition”. I don’t understand their views very well.)
Do you have references of posts of those people who think goal-directedness is binary-ish? That would be very useful, thanks. :)
Uh, not really. The mesa optimizers sequence sort of comes from this viewpoint, as does this question, but I haven’t really seen any posts arguing for this position.