Is it just me, or does this validate some of the parts of Yann LeCun’s “A Path Towards Autonomous Machine Intelligence” paper?
The two papers both use an algorithm consisting of multiple specialized models, with DreamerV3 using 3 models that seem very similar to those described by LeCun:
“the world model predicts future outcomes of potential actions”
“the critic judges the value of each situation”
“the actor learns to reach valuable situations”
World model, critic, actor—all are also described in LeCun’s paper. So are we seeing a successful push towards an AGI similar to LeCun’s ideas?
I’m not familiar with LeCun’s ideas, but I don’t think the idea of having an actor, critic, and world model is new in this paper. For a while, most RL algorithms have used an actor-critic architecture, including OpenAI’s old favorite PPO. Model-based RL has been around for years as well, so probably plenty of projects have used an actor, critic, and world model.
Even though the core idea isn’t novel, this paper getting good results might indicate that model-based RL is making more progress than expected, so if LeCun predicted that the future would look more like model-based RL, maybe he gets points for that.
Is it just me, or does this validate some of the parts of Yann LeCun’s “A Path Towards Autonomous Machine Intelligence” paper?
The two papers both use an algorithm consisting of multiple specialized models, with DreamerV3 using 3 models that seem very similar to those described by LeCun:
“the world model predicts future outcomes of potential actions”
“the critic judges the value of each situation”
“the actor learns to reach valuable situations”
World model, critic, actor—all are also described in LeCun’s paper. So are we seeing a successful push towards an AGI similar to LeCun’s ideas?
I’m not familiar with LeCun’s ideas, but I don’t think the idea of having an actor, critic, and world model is new in this paper. For a while, most RL algorithms have used an actor-critic architecture, including OpenAI’s old favorite PPO. Model-based RL has been around for years as well, so probably plenty of projects have used an actor, critic, and world model.
Even though the core idea isn’t novel, this paper getting good results might indicate that model-based RL is making more progress than expected, so if LeCun predicted that the future would look more like model-based RL, maybe he gets points for that.