This is fascinating, and is further evidence to me that LLMs contain models of reality.
I get frustrated with people who say LLMs “just” predict the next token, or they are simply copying and pasting bits of text from their training data. This argument skips over the fact that in order to accurately predict the next token, it’s necessary to compress the data in the training set down to something which looks a lot like a mostly accurate model of the world. In other words, if you have a large set of data entangled with reality, then the simplest model which predicts that data looks like reality.
This model of reality can be used to infer things which aren’t explicitly in the training data—like distances between places which aren’t mentioned together in the training data.
I’ve been thinking about this in the back of my mind for a while now. I think it lines up with points Cory Doctorow has made in talks about enshittification.
I’d like to see recommendation algorithms which are user-editable and preferably platform-agnostic, to allow low switching costs. A situation where people can build their own social media platform and install a recommendation algorithm which works for them, pulling in posts from other users across platforms who they follow. I’ve heard that the fediverse is trying to do something like this, but I’ve not been able to get engaged with it yet.
It’s cool to see efforts like Tournesol, though it’s a shame they don’t have a mobile extension yet.