Here’s another milestone in AI development that I expect to happen in the next few years which could be worth noting: I don’t think any of the large language models that currently exist write anything to an external memory. You can get a chatbot to hold a conversation and ‘remember’ what was said by appending the dialogue to its next input, but I’d imagine this would get unwieldy if you want your language model to keep track of details over a large number of interactions.
Fine-tuning a language model so that it makes use of a memory could lead to: 1. More consistent behavior 2. ‘Mesa-learning’ (it could learn things about the world from its inputs instead of just by gradient decent)
This seems relevant from a safety perspective because I can imagine ‘mesa-learning’ turning into ‘mesa-agency.’
Here’s another milestone in AI development that I expect to happen in the next few years which could be worth noting:
I don’t think any of the large language models that currently exist write anything to an external memory. You can get a chatbot to hold a conversation and ‘remember’ what was said by appending the dialogue to its next input, but I’d imagine this would get unwieldy if you want your language model to keep track of details over a large number of interactions.
Fine-tuning a language model so that it makes use of a memory could lead to:
1. More consistent behavior
2. ‘Mesa-learning’ (it could learn things about the world from its inputs instead of just by gradient decent)
This seems relevant from a safety perspective because I can imagine ‘mesa-learning’ turning into ‘mesa-agency.’