In other words, models are mostly misaligned because there are strong instrumental convergent incentives towards agency, and we don’t currently have any tools that allow us to shape the type of optimization that artificial systems are doing internally.
In the context of my comment, this appears to be an empirical claim about GPT-3. Is that right? (Otherwise I’m not sure what you are saying.)
If so, I don’t think this is right. On typical inputs I don’t think GPT-3 is instrumentally behaving well on the training distribution because it has a model fo the data-generating process.
I think on distribution you are mostly getting good behavior mostly either by not optimizing, or by optimizing for something we want. I think to the extent it’s malign it’s because there are possible inputs on which it is optimizing for something you don’t want, but those inputs are unlike those that appear in training and you have objective misgeneralization.
In that regime, I think the on-distribution performance is probably aligned and there is not much in-principle obstruction to using adversarial training to improve the robustness of alignment.
Instruct-GPT is not more aligned than GPT-3. It is more capable at performing many tasks, and we have some hope that some of the tasks at which it is getting better might help with AI Alignment down the lin
Could you define the word “alignment” as you are using it?
I’m using roughly the definition here. I think it’s the case that there are many inputs where GPT-3 is not trying to do what you want, but Instruct-GPT is. Indeed, I think Instruct-GPT is actually mostly trying to do what you want to the extent that it is trying to do anything at all. That would lead me to say it is more “aligned.”
I agree there are subtleties like “If I ask instruct-gpt to summarize a story, is it trying to summarize the story? Or trying to use that as evidence about ‘what Paul wants’ and then do that?” And I agree there is a real sense in which it isn’t smart enough for that distinction to be consistently meaningful, and so in that sense you might say my definition of intent alignment doesn’t really apply. (I more often think about models being “benign” or “malign,” more like asking: is it trying to optimize for something despite knowing that you wouldn’t like it.) I don’t think that’s what you are talking about here though.
right now, at the current state of the AI alignment field, the problem is not that we can’t provide good enough evaluation, or that we can only get good “average-case” performance, it’s that we have systems with random goals that are very far from human values or are capable of being reliably conservative.
If you have good oversight, I think you probably get good average case alignment. That’s ultimately an empirical claim about what happens when you do SGD, but the on-paper arguments looks quite good (namely: on-distribution alignment would improve the on-distribution performance and seems easy for SGD to learn relative to the complexity of the model itself) and it appears to match the data so far to the extent we have relevant data.
You seem to be confidently stating it’s false without engaging at all with the argument in favor or presenting or engaging with any empirical evidence.
You seem to be confidently stating it’s false without engaging at all with the argument in favor or presenting or engaging with any empirical evidence.
But which argument in favor did you present? You just said “the models are unaligned for these 2 reasons”, when those reasons do not seem comprehensive to me, and you did not give any justification for why those two reasons are comprehensive (or provide any links).
I tried to give a number of specific alternative reasons that do not seem to be covered by either of your two cases, and included a statement that we might disagree on definitional grounds, but that I don’t actually know what definitions you are using, and so can’t be confident that my critique makes sense.
Now that you’ve provided a definition, I still think what I said holds. My guess is there is a large inferential distance here, so I don’t think it makes sense to try to bridge that whole distance within this comment thread, though I will provide an additional round of responses.
If so, I don’t think this is right. On typical inputs I don’t think GPT-3 is instrumentally behaving well on the training distribution because it has a model fo the data-generating process.
I don’t think your definition of intent-alignment requires any unaligned system to have a model of the data-generating process, so I don’t understand the relevance of this. GPT-3 is not unaligned because it has a model of the data-generating process, and I didn’t claim that.
I did claim that neither GPT-3 nor Instruct-GPT are “trying to do what the operator wants it to do”, according to your definition, and that the primary reason for that is that in as much as its training process did produce a model that has “goals” and so can be modeled in any consequentialist terms, those “goals” do not match up with trying to be helpful to the operator. Most likely, they are a pretty messy objective we don’t really understand (which in the case of GPT-3 might be best described as “trying to generate text that in some simple latent space resembles the training distribution” and I don’t have any short description of what the “goals” of Instruct-GPT might be, though my guess is they are still pretty close to GPT-3s goals).
Indeed, I think Instruct-GPT is actually mostly trying to do what you want to the extent that it is trying to do anything at all. That would lead me to say it is more “aligned.”
I don’t think we know what Instruct-GPT is “trying to do”, and it seems unlikely to me that it is “trying to do what I want”. I agree in some sense it is “more trying to do what I want”, though not in a way that feels obviously very relevant to more capable systems, and not in a way that aligns very well with your intent definition (I feel like if I had to apply your linked definition to Instruct-GPT, I would say something like “ok, seems like it isn’t intent aligned, since the system doesn’t really seem to have much of an intent. And if there is a mechanism in its inner workings that corresponds to intent, we have no idea what thing it is pointed at, so probably it isn’t pointed at the right thing”).
And in either case, even if it is the case that if you squint your eyes a lot the system is “more aligned”, this doesn’t make the sentence “many of today’s systems are aligned unless humans make mistakes in evaluation or are deployed in high-stakes environments” true. “More aligned” is not equal to “aligned”.
The correct sentence seems to me “many of those systems are still mostly unaligned, but might be slightly more aligned than previous systems, though we have some hope that with better evaluation we can push that even further, and the misalignment problems are less bad on lower-stakes problems when we can rely on average-case performance, though overall the difference in alignment between GPT and Instruct-GPT is pretty unclear and probably not very large”.
I think on distribution you are mostly getting good behavior mostly either by not optimizing, or by optimizing for something we want. I think to the extent it’s malign it’s because there are possible inputs on which it is optimizing for something you don’t want, but those inputs are unlike those that appear in training and you have objective misgeneralization.
This seems wrong to me. On-distribution it seems to me that the system is usually optimizing for something that I don’t want. For example, GPT-3 primarily is trying to generate text that represents the distribution that its drawn from, which very rarely aligns with what I want (and is why prompt-engineering has such a large effect, e.g. “you are Albert Einstein” as a prefix improves performance on many tasks). Instruct-GPT does a bit better here, but probably most of its internal optimization power is still thrown at reasoning with the primary “intention” of generating text that is similar to its input distribution, since it seems unlikely that the fine-tuning completely rewrote most of these internal heuristics.
My guess is if Instruct-GPT was intent-aligned even for low-impact tasks, we could get it to be substantially more useful on many tasks. But my guess is what we currently have is mostly a model that is still primarily “trying” to generate text that is similar to its training distribution, with a few heuristics baked in in the human-feedback stage that make that text more likely to be a good fit for the question asked. In as much as the model is “trying to do something”, i.e. what most of its internal optimization power is pointed at, I am very skeptical that that is aligned with my task.
(Similarly, looking at Redwood’s recent model, it seems clear to me that they did not produce a model that “intents” to produce non-injurious completions. The model has two parts, one that is just “trying” to generate text similar to its training distribution, and a second part that is “trying” to detect whether a completion is injurious. This model seems clearly not intent-aligned, since almost none of its optimization power is going towards our target objective.)
If you have good oversight, I think you probably get good average case alignment. That’s ultimately an empirical claim about what happens when you do SGD, but the on-paper arguments looks quite good (namely: on-distribution alignment would improve the on-distribution performance and seems easy for SGD to learn relative to the complexity of the model itself) and it appears to match the data so far to the extent we have relevant data.
My guess is a lot of work is done here by the term “average case alignment”, so I am not fully sure how to respond. I disagree that the on-paper argument looks quite good, though it depends a lot on how narrowly you define “on-distribution”. Given my arguments above, you must either mean something different from intent-alignment (since to me at least it seems clear that Redwood’s model is not intent-aligned), or disagree with me on whether systems like Redwood’s are intent-aligned, in which case I don’t really know how to consistently apply your intent-alignment definition.
I also feel particularly confused about the term “average case alignment”, combined with “intent-alignment”. I can ascribe goals at multiple different levels to a model, and my guess is we both agree that describing current systems as having intentions at all is kind of fraught, but in as much as a model has a coherent goal, it seems like that goal is pretty consistent between different prompts, and so I am confused why we should expect average case alignment to be very different from normal alignment. It seems that if I have a model that is trying to do something, then asking it multiple times, probably won’t make a difference to its intention (I think, I mean, again, this all feels very handwavy, which is part of the reason why it feels so wrong to me to describe current models as “aligned”).
I currently think that the main relevant similarities between Instruct-GPT and a model that is trying to kill you, are about errors of the overseer (i.e. bad outputs to which they would give a high reward) or high-stakes errors (i.e. bad outputs which can have catastrophic effects before they are corrected by fine-tuning).
I’m interested in other kinds of relevant similarities, since I think those would be exciting and productive things to research. I don’t think the framework “Instruct-GPT and GPT-3 e.g. copy patterns that they saw in the prompt, so they are ‘trying’ to predict the next word and hence are misaligned” is super useful, though I see where it’s coming from and agree that I started it by using the word “aligned”.
Relatedly, and contrary to my original comment, I do agree that there can be bad intentional behavior left over from pre-training. This is a big part what ML researchers are motivated by when they talk about improving the sample-efficiency of RLHF. I usually try to discourage people from working on this issue, because it seems like something that will predictably get better rather than worse as models improve (and I expect you are even less happy with it than I am).
I agree that there is a lot of inferential distance, and it doesn’t seem worth trying to close the gap here. I’ve tried to write down a fair amount about my views, and I’m always interested to read arguments / evidence / intuitions for more pessimistic conclusions.
Similarly, looking at Redwood’s recent model, it seems clear to me that they did not produce a model that “intents” to produce non-injurious completions.
I agree with this, though it’s unrelated to the stated motivation for that project or to its relationship to long-term risk.
I currently think that the main relevant similarities between Instruct-GPT and a model that is trying to kill you, are about errors of the overseer (i.e. bad outputs to which they would give a high reward) or high-stakes errors (i.e. bad outputs which can have catastrophic effects before they are corrected by fine-tuning).
Phrased this way, I still disagree, but I think I disagree less strongly, and feel less of a need to respond to this. I care particularly much about using terms like “aligned” in consistent ways. Importantly, having powerful intent-aligned systems is much more useful than having powerful systems that just fail to kill you (e.g. because they are very conservative), and so getting to powerful aligned systems is a win-condition in the way that getting to powerful non-catastrophic systems is not.
I agree with this, though it’s unrelated to the stated motivation for that project or to its relationship to long-term risk.
Yep, I didn’t intend to imply that this was in contrast to the intention of the research. It was just on my mind as a recent architecture that I was confident we both had thought about, and so could use as a convenient example.
In the context of my comment, this appears to be an empirical claim about GPT-3. Is that right? (Otherwise I’m not sure what you are saying.)
If so, I don’t think this is right. On typical inputs I don’t think GPT-3 is instrumentally behaving well on the training distribution because it has a model fo the data-generating process.
I think on distribution you are mostly getting good behavior mostly either by not optimizing, or by optimizing for something we want. I think to the extent it’s malign it’s because there are possible inputs on which it is optimizing for something you don’t want, but those inputs are unlike those that appear in training and you have objective misgeneralization.
In that regime, I think the on-distribution performance is probably aligned and there is not much in-principle obstruction to using adversarial training to improve the robustness of alignment.
Could you define the word “alignment” as you are using it?
I’m using roughly the definition here. I think it’s the case that there are many inputs where GPT-3 is not trying to do what you want, but Instruct-GPT is. Indeed, I think Instruct-GPT is actually mostly trying to do what you want to the extent that it is trying to do anything at all. That would lead me to say it is more “aligned.”
I agree there are subtleties like “If I ask instruct-gpt to summarize a story, is it trying to summarize the story? Or trying to use that as evidence about ‘what Paul wants’ and then do that?” And I agree there is a real sense in which it isn’t smart enough for that distinction to be consistently meaningful, and so in that sense you might say my definition of intent alignment doesn’t really apply. (I more often think about models being “benign” or “malign,” more like asking: is it trying to optimize for something despite knowing that you wouldn’t like it.) I don’t think that’s what you are talking about here though.
If you have good oversight, I think you probably get good average case alignment. That’s ultimately an empirical claim about what happens when you do SGD, but the on-paper arguments looks quite good (namely: on-distribution alignment would improve the on-distribution performance and seems easy for SGD to learn relative to the complexity of the model itself) and it appears to match the data so far to the extent we have relevant data.
You seem to be confidently stating it’s false without engaging at all with the argument in favor or presenting or engaging with any empirical evidence.
But which argument in favor did you present? You just said “the models are unaligned for these 2 reasons”, when those reasons do not seem comprehensive to me, and you did not give any justification for why those two reasons are comprehensive (or provide any links).
I tried to give a number of specific alternative reasons that do not seem to be covered by either of your two cases, and included a statement that we might disagree on definitional grounds, but that I don’t actually know what definitions you are using, and so can’t be confident that my critique makes sense.
Now that you’ve provided a definition, I still think what I said holds. My guess is there is a large inferential distance here, so I don’t think it makes sense to try to bridge that whole distance within this comment thread, though I will provide an additional round of responses.
I don’t think your definition of intent-alignment requires any unaligned system to have a model of the data-generating process, so I don’t understand the relevance of this. GPT-3 is not unaligned because it has a model of the data-generating process, and I didn’t claim that.
I did claim that neither GPT-3 nor Instruct-GPT are “trying to do what the operator wants it to do”, according to your definition, and that the primary reason for that is that in as much as its training process did produce a model that has “goals” and so can be modeled in any consequentialist terms, those “goals” do not match up with trying to be helpful to the operator. Most likely, they are a pretty messy objective we don’t really understand (which in the case of GPT-3 might be best described as “trying to generate text that in some simple latent space resembles the training distribution” and I don’t have any short description of what the “goals” of Instruct-GPT might be, though my guess is they are still pretty close to GPT-3s goals).
I don’t think we know what Instruct-GPT is “trying to do”, and it seems unlikely to me that it is “trying to do what I want”. I agree in some sense it is “more trying to do what I want”, though not in a way that feels obviously very relevant to more capable systems, and not in a way that aligns very well with your intent definition (I feel like if I had to apply your linked definition to Instruct-GPT, I would say something like “ok, seems like it isn’t intent aligned, since the system doesn’t really seem to have much of an intent. And if there is a mechanism in its inner workings that corresponds to intent, we have no idea what thing it is pointed at, so probably it isn’t pointed at the right thing”).
And in either case, even if it is the case that if you squint your eyes a lot the system is “more aligned”, this doesn’t make the sentence “many of today’s systems are aligned unless humans make mistakes in evaluation or are deployed in high-stakes environments” true. “More aligned” is not equal to “aligned”.
The correct sentence seems to me “many of those systems are still mostly unaligned, but might be slightly more aligned than previous systems, though we have some hope that with better evaluation we can push that even further, and the misalignment problems are less bad on lower-stakes problems when we can rely on average-case performance, though overall the difference in alignment between GPT and Instruct-GPT is pretty unclear and probably not very large”.
This seems wrong to me. On-distribution it seems to me that the system is usually optimizing for something that I don’t want. For example, GPT-3 primarily is trying to generate text that represents the distribution that its drawn from, which very rarely aligns with what I want (and is why prompt-engineering has such a large effect, e.g. “you are Albert Einstein” as a prefix improves performance on many tasks). Instruct-GPT does a bit better here, but probably most of its internal optimization power is still thrown at reasoning with the primary “intention” of generating text that is similar to its input distribution, since it seems unlikely that the fine-tuning completely rewrote most of these internal heuristics.
My guess is if Instruct-GPT was intent-aligned even for low-impact tasks, we could get it to be substantially more useful on many tasks. But my guess is what we currently have is mostly a model that is still primarily “trying” to generate text that is similar to its training distribution, with a few heuristics baked in in the human-feedback stage that make that text more likely to be a good fit for the question asked. In as much as the model is “trying to do something”, i.e. what most of its internal optimization power is pointed at, I am very skeptical that that is aligned with my task.
(Similarly, looking at Redwood’s recent model, it seems clear to me that they did not produce a model that “intents” to produce non-injurious completions. The model has two parts, one that is just “trying” to generate text similar to its training distribution, and a second part that is “trying” to detect whether a completion is injurious. This model seems clearly not intent-aligned, since almost none of its optimization power is going towards our target objective.)
My guess is a lot of work is done here by the term “average case alignment”, so I am not fully sure how to respond. I disagree that the on-paper argument looks quite good, though it depends a lot on how narrowly you define “on-distribution”. Given my arguments above, you must either mean something different from intent-alignment (since to me at least it seems clear that Redwood’s model is not intent-aligned), or disagree with me on whether systems like Redwood’s are intent-aligned, in which case I don’t really know how to consistently apply your intent-alignment definition.
I also feel particularly confused about the term “average case alignment”, combined with “intent-alignment”. I can ascribe goals at multiple different levels to a model, and my guess is we both agree that describing current systems as having intentions at all is kind of fraught, but in as much as a model has a coherent goal, it seems like that goal is pretty consistent between different prompts, and so I am confused why we should expect average case alignment to be very different from normal alignment. It seems that if I have a model that is trying to do something, then asking it multiple times, probably won’t make a difference to its intention (I think, I mean, again, this all feels very handwavy, which is part of the reason why it feels so wrong to me to describe current models as “aligned”).
I currently think that the main relevant similarities between Instruct-GPT and a model that is trying to kill you, are about errors of the overseer (i.e. bad outputs to which they would give a high reward) or high-stakes errors (i.e. bad outputs which can have catastrophic effects before they are corrected by fine-tuning).
I’m interested in other kinds of relevant similarities, since I think those would be exciting and productive things to research. I don’t think the framework “Instruct-GPT and GPT-3 e.g. copy patterns that they saw in the prompt, so they are ‘trying’ to predict the next word and hence are misaligned” is super useful, though I see where it’s coming from and agree that I started it by using the word “aligned”.
Relatedly, and contrary to my original comment, I do agree that there can be bad intentional behavior left over from pre-training. This is a big part what ML researchers are motivated by when they talk about improving the sample-efficiency of RLHF. I usually try to discourage people from working on this issue, because it seems like something that will predictably get better rather than worse as models improve (and I expect you are even less happy with it than I am).
I agree that there is a lot of inferential distance, and it doesn’t seem worth trying to close the gap here. I’ve tried to write down a fair amount about my views, and I’m always interested to read arguments / evidence / intuitions for more pessimistic conclusions.
I agree with this, though it’s unrelated to the stated motivation for that project or to its relationship to long-term risk.
Phrased this way, I still disagree, but I think I disagree less strongly, and feel less of a need to respond to this. I care particularly much about using terms like “aligned” in consistent ways. Importantly, having powerful intent-aligned systems is much more useful than having powerful systems that just fail to kill you (e.g. because they are very conservative), and so getting to powerful aligned systems is a win-condition in the way that getting to powerful non-catastrophic systems is not.
Yep, I didn’t intend to imply that this was in contrast to the intention of the research. It was just on my mind as a recent architecture that I was confident we both had thought about, and so could use as a convenient example.