One-line version of this post: What do Wittgensteinian language games and NLP word embeddings have in common?
Four-line version of this post: Relational, praxis-based connections between concepts, represented as “distances” in multidimensional space, capture meaning. The shorter the distance, the more related the concepts. This is how Word2vec works, what Wittgenstein was describing with “language games,” and also the way cell biologists are analyzing the peripheral blood these days. Are these relational maps the way to think about thinking?
Multi-line version of this post: This is my first post on LessWrong. (Hi!) I’d love to be less wrong about it.
I was sitting in a meeting that was 50% biologists and 50% computer scientists. The topic of the meeting was about ways to process multi-parametric datasets, where each cell in the peripheral blood was tagged by multiple surface markers that related back to its phenotype and therefore its identity. (The algorithm in question was t-Distributed Stochastic Neighbor Embedding.) Immunologists used to think a T-cell was a T-cell. But in that meeting, we were considering a smear of T-cells in a 32-dimensional T-cell space, clustered by their properties and functional status (activated or exhausted; killer or memory etc).
In the meeting, as I was looking at colored 2D and 3D representations that abstracted features of that higher dimensional space (activated killer T cells on the bottom left in blue; resting memory cells on top in orange; what’s that weird purple cluster in the bottom left? and so on), it occurred to me that this technique was probably good at capturing meaning across the board.
Abstracting meaning from measured distances between mapped concepts isn’t a new idea. It’s described beautifully in The Cluster Structure of Thingspace. I just wonder if we can ride it a little further into the fog.
Wittgenstein is often quoted in the Venn diagram overlap region between meaning and computation. The strongest applicable Wittgensteinian concept to this particular space is his idea of a language game. A language game is a process in which words are used according to specific rules and contexts, shaping how we understand meaning and communicate. LessWrong has discussions on the relationship between language games and truth, such as in Parasitic Language Games: maintaining ambiguity to hide conflict while burning the commons, but searching the site reveals less content directly connecting Wittgenstein to phase space, vector space, or thingspace than I’d expect.
Clustering of things in thingspace isn’t a direct Wittgensteinian language game (I don’t think). It seems more like what you’d get if you took a Wittgensteinian approach (praxis-based, relational) and used it to build a vector space for topologies of concept (i.e. for “chairness” and “birdness” and “Golden Gate Bridgeness”).
Word2vec, a natural language processing model, does a simple form of this when it represents words with similar meanings close together in vector space. It seems LLMs do a version of this, with Golden Gate Claude supporting the idea that within LLMs concepts can be topologically localized.
I don’t think there’s enough understood about language processing in the brain to say with certainty that the brain also clusters concepts like this, but I’m guessing it’s quite likely.
Short distances between conceptual nodes in a vast relational web seems like a good way to convey meaning. It works for an understanding of concrete words and literal T-cell properties, but it’s also a relational process that maps back to more abstract concepts. In a way, traversing such maps, building patterns within them, running patterns through them, operating on the data they contain, is probably the best operational definition of “thinking” that I can think of.
Wittgenstein and Word2vec: Capturing Relational Meaning in Language and Thought
One-line version of this post: What do Wittgensteinian language games and NLP word embeddings have in common?
Four-line version of this post: Relational, praxis-based connections between concepts, represented as “distances” in multidimensional space, capture meaning. The shorter the distance, the more related the concepts. This is how Word2vec works, what Wittgenstein was describing with “language games,” and also the way cell biologists are analyzing the peripheral blood these days. Are these relational maps the way to think about thinking?
Multi-line version of this post: This is my first post on LessWrong. (Hi!) I’d love to be less wrong about it.
I was sitting in a meeting that was 50% biologists and 50% computer scientists. The topic of the meeting was about ways to process multi-parametric datasets, where each cell in the peripheral blood was tagged by multiple surface markers that related back to its phenotype and therefore its identity. (The algorithm in question was t-Distributed Stochastic Neighbor Embedding.) Immunologists used to think a T-cell was a T-cell. But in that meeting, we were considering a smear of T-cells in a 32-dimensional T-cell space, clustered by their properties and functional status (activated or exhausted; killer or memory etc).
In the meeting, as I was looking at colored 2D and 3D representations that abstracted features of that higher dimensional space (activated killer T cells on the bottom left in blue; resting memory cells on top in orange; what’s that weird purple cluster in the bottom left? and so on), it occurred to me that this technique was probably good at capturing meaning across the board.
Abstracting meaning from measured distances between mapped concepts isn’t a new idea. It’s described beautifully in The Cluster Structure of Thingspace. I just wonder if we can ride it a little further into the fog.
Wittgenstein is often quoted in the Venn diagram overlap region between meaning and computation. The strongest applicable Wittgensteinian concept to this particular space is his idea of a language game. A language game is a process in which words are used according to specific rules and contexts, shaping how we understand meaning and communicate. LessWrong has discussions on the relationship between language games and truth, such as in Parasitic Language Games: maintaining ambiguity to hide conflict while burning the commons, but searching the site reveals less content directly connecting Wittgenstein to phase space, vector space, or thingspace than I’d expect.
Clustering of things in thingspace isn’t a direct Wittgensteinian language game (I don’t think). It seems more like what you’d get if you took a Wittgensteinian approach (praxis-based, relational) and used it to build a vector space for topologies of concept (i.e. for “chairness” and “birdness” and “Golden Gate Bridgeness”).
Word2vec, a natural language processing model, does a simple form of this when it represents words with similar meanings close together in vector space. It seems LLMs do a version of this, with Golden Gate Claude supporting the idea that within LLMs concepts can be topologically localized.
I don’t think there’s enough understood about language processing in the brain to say with certainty that the brain also clusters concepts like this, but I’m guessing it’s quite likely.
Short distances between conceptual nodes in a vast relational web seems like a good way to convey meaning. It works for an understanding of concrete words and literal T-cell properties, but it’s also a relational process that maps back to more abstract concepts. In a way, traversing such maps, building patterns within them, running patterns through them, operating on the data they contain, is probably the best operational definition of “thinking” that I can think of.
…Thoughts?