It seems like this would be because the transformer weights are fixed and we have not built a mechanism for the model to record things it needs to learn to improve performance or an automated way to practice offline to do so.
It’s just missing all this, like a human patient with large sections of their brain surgically removed. Doesn’t seem difficult or long term to add this does it? How many years before one of the competing AI lab adds some form of “performance enhancing fine tuning and self play”?
It seems like this would be because the transformer weights are fixed and we have not built a mechanism for the model to record things it needs to learn to improve performance or an automated way to practice offline to do so.
It’s just missing all this, like a human patient with large sections of their brain surgically removed. Doesn’t seem difficult or long term to add this does it? How many years before one of the competing AI lab adds some form of “performance enhancing fine tuning and self play”?
Less than a year. They probably already have toy models with periodically or continuously updating weights.