The MIRI 2000s paradigm for an AI capable of self-improvement, was that it would be modular code with a hierarchical organization, that would potentially engage in self-improvement at every level.
The actual path we’ve been on has been: deep learning, scaling, finetuning with RLHF, and now (just starting) reflective agents built on a GPT base.
A reflective GPT-based agent is certainly capable of studying itself and coming up with ideas for improvement. So we’re probably at the beginning of attempts at self-improvement, right now.
The MIRI 2000s paradigm for an AI capable of self-improvement, was that it would be modular code with a hierarchical organization, that would potentially engage in self-improvement at every level.
The actual path we’ve been on has been: deep learning, scaling, finetuning with RLHF, and now (just starting) reflective agents built on a GPT base.
A reflective GPT-based agent is certainly capable of studying itself and coming up with ideas for improvement. So we’re probably at the beginning of attempts at self-improvement, right now.