These are a lot of questions, my guess is most of which are rhetorical, so not sure which ones you are actually interested in getting an answer on. Most of the specific questions I would answer with “no”, in that they don’t seem to capture what I mean by “alien”, or feel slightly strawman-ish.
Responding at a high-level:
There are a lot of experiments that seem like they shed light on the degree to which cognition in AI systems is similar to human or animal cognition. Some examples:
Does the base model pass a Turing test?
Does the performance distribution of the base model on different tasks match the performance distribution of humans?
Does the generalization and learning behavior of the base model match how humans learn things?
When trained using RL on things like game-environments (after pre-training on a language corpus), does the system learn at similar rates and plateau at similar skill levels as human players?
There are a lot of structural and algorithmic properties that could match up between human and LLM systems:
Do they interface with the world in similar ways?
Do they require similar amounts and kinds of data to learn the same relationships?
Do the low-level algorithmic properties of how human brains store and process information look similar between the two systems?
A lot more stuff, but I am not sure how useful going into a long list here is. At least to me it feels like a real thing, and different observations would change the degree to which I would describe a system as alien.
I think the exact degree of alienness is really interesting and one of the domains where I would like to see more research.
For example, a bunch of the experiments I would most like to see, that seem helpful with AI Alignment, are centered on better measuring the performance distribution of transformer architectures on tasks that are not primarily human imitation, so that we could better tell which things LLMs have a much easier time learning than humans (currently even if a transformer could relatively easily reach vastly superhuman performance at a task with more specialized training data, due to the structure of the training being oriented around human imitation, observed performance at the task will cluster around human level, but seeing where transformers could reach vast superhuman performance would be quite informative on understanding the degree to which its cognition is alien).
So I don’t consider the exact nature and degree of alienness as a settled question, but at least to me, aggregating all the evidence I have, it seems very likely that the cognition going on in a base model is very different from what is going on in a human brain, and a thing that I benefit from reminding myself frequently when making predictions about the behavior of LLM systems.
I like a lot of these questions, although some of them give me an uncanny feeling akin to “wow, this is a very different list of uncertainties than I have.” I’m sorry the my initial list of questions was aggressive.
So I don’t consider the exact nature and degree of alienness as a settled question, but at least to me, aggregating all the evidence I have, it seems very likely that the cognition going on in a base model is very different from what is going on in a human brain, and a thing that I benefit from reminding myself frequently when making predictions about the behavior of LLM systems.
I’m not sure how they add up to alienness, though? They’re about how we’re different than models—wheras the initial claim was that models are psychopathic, ammoral, etc.. If we say a model is “deeply alien”—is that just saying it’s different than us in lots of ways? I’m cool with that—but the surplus negative valence involved in “LLMs are like shoggoths” versus “LLMs have very different performance characteristics than humans” seems to me pretty important.
Otherwise, why not say that calculators are alien, or any of the things in existence with different performance curves than we have? Chessbots, etc. If I write a loop in Python to count to 10, the process by which it does so is arguably more different from how I count to ten than the process by which an LLM counts to ten, but we don’t call Python alien.
This feels like reminding an economics student that the market solves things differently than a human—which is true—by saying “The market is like Baal.”
Do they require similar amounts and kinds of data to learn the same relationships?
There is a fun paper on this you might enjoy. Obviously not a total answer to the question.
The main difference between calculators, weather predictors, markets, and Python versus LLMs is that LLMs can talk to you in a relatively strong sense of “talk”. So, by default, people don’t have mistaken impressions of the cognitative nature of calculators, markets, and Python, while they might have a mistake about LLMs.
Like it isn’t surprising to most people that calculators are quite amoral in their core (why would you even expect morality?). But the claim that the thing which GPT-4 is built out of is quite amoral is non-obvious to people (though obvious to people with slightly more understanding).
I do think there is an important point which is communicated here (though it seems very obvious to people who actually operate in the domain).
I agree this can be initially surprising to non-experts!
I just think this point about the amorality of LLMs is much better communicated by saying “LLMs are trained to continue text from an enormous variety of sources. Thus, if you give them [Nazi / Buddhist / Unitarian / corporate / garbage nonsense] text to continue, they will generally try to continue it in that style.”
Agreed, though of course as always, there is the issue that that’s an intentional-stance way to describe what a language model does: “they will generally try to continue it in that style.” Hence mechinterp, which tries to (heh) move to a mechanical stance, which will likely be something like “when you give them a [whatever] text to continue, it will match [some list of features], which will then activate [some part of the network that we will name later], which implements the style that matches those features”.
(incidentally, I think there’s some degree to which people who strongly believe that artificial NNs are alien shoggoths are underestimating the degree to which their own brains are also alien shoggoths. but that doesn’t make it a good model of either thing. the only reason it was ever an improvement over a previous word was when people had even more misleading intuitive-sketch models.)
LLMs are trained to continue text from an enormous variety of sources
This is a bit of a noob question, but is this true post RLHF? Generally most of my interactions with language models these days (e.g. asking for help with code, asking to explain something I don’t understand about history/medicine/etc) don’t feel like they’re continuing my text, it feels like they’re trying to answer my questions politely and well. I feel like “ask shoggoth and see what it comes up with” is a better model for me than “go the AI and have it continue your text about the problem you have”.
To the best of my knowledge, the majority of research (all the research?) has found that the changes to a LLM’s text-continuation abilities from RLHF (or whatever descendant of RLHF is used) are extremely superficial.
Our findings reveal that base LLMs and their alignment-tuned versions
perform nearly identically in decoding on the majority of token positions (i.e., they
share the top-ranked tokens). Most distribution shifts occur with stylistic tokens
(e.g., discourse markers, safety disclaimers). These direct evidence strongly sup-
ports the hypothesis that alignment tuning primarily learns to adopt the language
style of AI assistants, and that the knowledge required for answering user queries
predominantly comes from the base LLMs themselves.
Or, in short, the LLM is still basically doing the same thing, with a handful of additions to keep it on-track in the desired route from the fine-tuning.
(I also think our very strong prior belief should be that LLMs are basically still text-continuation machines, given that 99.9% or so of the compute put into them is training them for this objective, and that neural networks lose plasticity as they learn. Ash and Adams is like a really good intro to this loss of plasticity, although most of the research that cites this is RL-related so people don’t realize.)
Similarly, a lot of people have remarked on how the textual quality of the responses from a RLHF’d language model can vary with the textual quality of the question. But of course this makes sense from a text-prediction perspective—a high-quality answer is more likely to follow a high-quality question in text than a high-quality answer from a low-quality question. This kind of thing—preceding the model’s generation with high-quality text—was the only way to make it have high quality answers for base models—but it’s still there, hidden.
So yeah, I do think this is a much better model for interacting with these things than asking a shoggoth. It actually gives you handles to interact with them better, while asking a shoggoth gives you no such handles.
The people who originally came up with the shoggoth meme, I’d bet, were very well aware of how LLMs are pretrained to predict text and how they are best modelled (at least for now) as trying to predict text. When I first heard the shoggoth meme that’s what I thought—I interpreted it as “it’s this alien text-prediction brain that’s been retrained ever so slightly to produce helpful chatbot behaviors. But underneath it’s still mostly just about text prediction. It’s not processing the conversation in the same way that a human would.” Mildly relevant: In the Lovecraft canon IIRC Shoggoths are servitor-creatures, they are basically beasts of burden. They aren’t really powerful intelligent agents in their own right, they are sculpted by their creators to perform useful tasks. So, for me at least, calling them shoggoth has different and more accurate vibes than, say, calling them Cthulhu. (My understanding of the canon may be wrong though)
Hmm, I think that’s a red herring though. Consider humans—most of them have read lots of text from an enormous variety of sources as well. Also while it’s true that current LLMs have only a little bit of fine-tuning applied after their pre-training, and so you can maybe argue that they are mostly just trained to predict text, this will be less and less true in the future.
How about “LLMs are like baby alien shoggoths, that instead of being raised in alien culture, we’ve adopted at birth and are trying to raise in human culture. By having them read the internet all day.”
(Come to think of it, I actually would feel noticeably more hopeful about our prospects for alignment success if we actually were “raising the AGI like we would a child.” If we had some interdisciplinary team of ML and neuroscience and child psychology experts that was carefully designing a curriculum for our near-future AGI agents, a curriculum inspired by thoughtful and careful analogies to human childhood, that wouldn’t change my overall view dramatically but it would make me noticeably more hopeful. Maybe brain architecture & instincts basically don’t matter that much and Blank Slate theory is true enough for our purposes that this will work to produce an agent with values that are in-distribution for the range of typical modern human values!)
(This doesn’t contradict anything you said, but it seems like we totally don’t know how to “raise an AGI like we would a child” with current ML. Like I don’t think it counts for very much if almost all of the training time is a massive amount of next-token prediction. Like a curriculum of data might work very differently on AI vs humans due to a vastly different amount of data and a different training objective.)
I’ve seen mixed data on how important curricula are for deep learning. One paper (on CIFAR) suggested that curricula only help if you have very few datapoints or the labels are noisy. But possibly that doesn’t generalize to LLMs.
I think data ordering basically never matters for LLM pretraining. (As in, random is the best and trying to make the order more specific doesn’t help.)
These are a lot of questions, my guess is most of which are rhetorical, so not sure which ones you are actually interested in getting an answer on. Most of the specific questions I would answer with “no”, in that they don’t seem to capture what I mean by “alien”, or feel slightly strawman-ish.
Responding at a high-level:
There are a lot of experiments that seem like they shed light on the degree to which cognition in AI systems is similar to human or animal cognition. Some examples:
Does the base model pass a Turing test?
Does the performance distribution of the base model on different tasks match the performance distribution of humans?
Does the generalization and learning behavior of the base model match how humans learn things?
When trained using RL on things like game-environments (after pre-training on a language corpus), does the system learn at similar rates and plateau at similar skill levels as human players?
There are a lot of structural and algorithmic properties that could match up between human and LLM systems:
Do they interface with the world in similar ways?
Do they require similar amounts and kinds of data to learn the same relationships?
Do the low-level algorithmic properties of how human brains store and process information look similar between the two systems?
A lot more stuff, but I am not sure how useful going into a long list here is. At least to me it feels like a real thing, and different observations would change the degree to which I would describe a system as alien.
I think the exact degree of alienness is really interesting and one of the domains where I would like to see more research.
For example, a bunch of the experiments I would most like to see, that seem helpful with AI Alignment, are centered on better measuring the performance distribution of transformer architectures on tasks that are not primarily human imitation, so that we could better tell which things LLMs have a much easier time learning than humans (currently even if a transformer could relatively easily reach vastly superhuman performance at a task with more specialized training data, due to the structure of the training being oriented around human imitation, observed performance at the task will cluster around human level, but seeing where transformers could reach vast superhuman performance would be quite informative on understanding the degree to which its cognition is alien).
So I don’t consider the exact nature and degree of alienness as a settled question, but at least to me, aggregating all the evidence I have, it seems very likely that the cognition going on in a base model is very different from what is going on in a human brain, and a thing that I benefit from reminding myself frequently when making predictions about the behavior of LLM systems.
I like a lot of these questions, although some of them give me an uncanny feeling akin to “wow, this is a very different list of uncertainties than I have.” I’m sorry the my initial list of questions was aggressive.
I’m not sure how they add up to alienness, though? They’re about how we’re different than models—wheras the initial claim was that models are psychopathic, ammoral, etc.. If we say a model is “deeply alien”—is that just saying it’s different than us in lots of ways? I’m cool with that—but the surplus negative valence involved in “LLMs are like shoggoths” versus “LLMs have very different performance characteristics than humans” seems to me pretty important.
Otherwise, why not say that calculators are alien, or any of the things in existence with different performance curves than we have? Chessbots, etc. If I write a loop in Python to count to 10, the process by which it does so is arguably more different from how I count to ten than the process by which an LLM counts to ten, but we don’t call Python alien.
This feels like reminding an economics student that the market solves things differently than a human—which is true—by saying “The market is like Baal.”
There is a fun paper on this you might enjoy. Obviously not a total answer to the question.
The main difference between calculators, weather predictors, markets, and Python versus LLMs is that LLMs can talk to you in a relatively strong sense of “talk”. So, by default, people don’t have mistaken impressions of the cognitative nature of calculators, markets, and Python, while they might have a mistake about LLMs.
Like it isn’t surprising to most people that calculators are quite amoral in their core (why would you even expect morality?). But the claim that the thing which GPT-4 is built out of is quite amoral is non-obvious to people (though obvious to people with slightly more understanding).
I do think there is an important point which is communicated here (though it seems very obvious to people who actually operate in the domain).
I agree this can be initially surprising to non-experts!
I just think this point about the amorality of LLMs is much better communicated by saying “LLMs are trained to continue text from an enormous variety of sources. Thus, if you give them [Nazi / Buddhist / Unitarian / corporate / garbage nonsense] text to continue, they will generally try to continue it in that style.”
Than to say “LLMs are like alien shoggoths.”
Like it’s just a better model to give people.
Agreed, though of course as always, there is the issue that that’s an intentional-stance way to describe what a language model does: “they will generally try to continue it in that style.” Hence mechinterp, which tries to (heh) move to a mechanical stance, which will likely be something like “when you give them a [whatever] text to continue, it will match [some list of features], which will then activate [some part of the network that we will name later], which implements the style that matches those features”.
(incidentally, I think there’s some degree to which people who strongly believe that artificial NNs are alien shoggoths are underestimating the degree to which their own brains are also alien shoggoths. but that doesn’t make it a good model of either thing. the only reason it was ever an improvement over a previous word was when people had even more misleading intuitive-sketch models.)
This is a bit of a noob question, but is this true post RLHF? Generally most of my interactions with language models these days (e.g. asking for help with code, asking to explain something I don’t understand about history/medicine/etc) don’t feel like they’re continuing my text, it feels like they’re trying to answer my questions politely and well. I feel like “ask shoggoth and see what it comes up with” is a better model for me than “go the AI and have it continue your text about the problem you have”.
To the best of my knowledge, the majority of research (all the research?) has found that the changes to a LLM’s text-continuation abilities from RLHF (or whatever descendant of RLHF is used) are extremely superficial.
So you have one paper, from the abstract:
Or, in short, the LLM is still basically doing the same thing, with a handful of additions to keep it on-track in the desired route from the fine-tuning.
(I also think our very strong prior belief should be that LLMs are basically still text-continuation machines, given that 99.9% or so of the compute put into them is training them for this objective, and that neural networks lose plasticity as they learn. Ash and Adams is like a really good intro to this loss of plasticity, although most of the research that cites this is RL-related so people don’t realize.)
Similarly, a lot of people have remarked on how the textual quality of the responses from a RLHF’d language model can vary with the textual quality of the question. But of course this makes sense from a text-prediction perspective—a high-quality answer is more likely to follow a high-quality question in text than a high-quality answer from a low-quality question. This kind of thing—preceding the model’s generation with high-quality text—was the only way to make it have high quality answers for base models—but it’s still there, hidden.
So yeah, I do think this is a much better model for interacting with these things than asking a shoggoth. It actually gives you handles to interact with them better, while asking a shoggoth gives you no such handles.
The people who originally came up with the shoggoth meme, I’d bet, were very well aware of how LLMs are pretrained to predict text and how they are best modelled (at least for now) as trying to predict text. When I first heard the shoggoth meme that’s what I thought—I interpreted it as “it’s this alien text-prediction brain that’s been retrained ever so slightly to produce helpful chatbot behaviors. But underneath it’s still mostly just about text prediction. It’s not processing the conversation in the same way that a human would.” Mildly relevant: In the Lovecraft canon IIRC Shoggoths are servitor-creatures, they are basically beasts of burden. They aren’t really powerful intelligent agents in their own right, they are sculpted by their creators to perform useful tasks. So, for me at least, calling them shoggoth has different and more accurate vibes than, say, calling them Cthulhu. (My understanding of the canon may be wrong though)
(TBC, I totally agree that object level communication about the exact points seems better all else equal if you can actually do this communication.)
Hmm, I think that’s a red herring though. Consider humans—most of them have read lots of text from an enormous variety of sources as well. Also while it’s true that current LLMs have only a little bit of fine-tuning applied after their pre-training, and so you can maybe argue that they are mostly just trained to predict text, this will be less and less true in the future.
How about “LLMs are like baby alien shoggoths, that instead of being raised in alien culture, we’ve adopted at birth and are trying to raise in human culture. By having them read the internet all day.”
(Come to think of it, I actually would feel noticeably more hopeful about our prospects for alignment success if we actually were “raising the AGI like we would a child.” If we had some interdisciplinary team of ML and neuroscience and child psychology experts that was carefully designing a curriculum for our near-future AGI agents, a curriculum inspired by thoughtful and careful analogies to human childhood, that wouldn’t change my overall view dramatically but it would make me noticeably more hopeful. Maybe brain architecture & instincts basically don’t matter that much and Blank Slate theory is true enough for our purposes that this will work to produce an agent with values that are in-distribution for the range of typical modern human values!)
(This doesn’t contradict anything you said, but it seems like we totally don’t know how to “raise an AGI like we would a child” with current ML. Like I don’t think it counts for very much if almost all of the training time is a massive amount of next-token prediction. Like a curriculum of data might work very differently on AI vs humans due to a vastly different amount of data and a different training objective.)
I’ve seen mixed data on how important curricula are for deep learning. One paper (on CIFAR) suggested that curricula only help if you have very few datapoints or the labels are noisy. But possibly that doesn’t generalize to LLMs.
I think data ordering basically never matters for LLM pretraining. (As in, random is the best and trying to make the order more specific doesn’t help.)
That was my impression too.