With the current transformer models we see that once a model is trained not only direct copies of it are created but also derivates that are smaller and potentially trained to be able to be better at a task.
Just like human cognitive diversity is useful to act in the world it’s likely also more effective to have slight divergence in AGI models.
With the current transformer models we see that once a model is trained not only direct copies of it are created but also derivates that are smaller and potentially trained to be able to be better at a task.
Just like human cognitive diversity is useful to act in the world it’s likely also more effective to have slight divergence in AGI models.