Taboo “Outside View”
No one has ever seen an AGI takeoff, so any attempt to understand it must use these outside view considerations.
—[Redacted for privacy]
What? That’s exactly backwards. If we had lots of experience with past AGI takeoffs, using the outside view to predict the next one would be a lot more effective.
—My reaction
Two years ago I wrote a deep-dive summary of Superforecasting and the associated scientific literature. I learned about the “Outside view” / “Inside view” distinction, and the evidence supporting it. At the time I was excited about the concept and wrote: “...I think we should do our best to imitate these best-practices, and that means using the outside view far more than we would naturally be inclined.”
Now that I have more experience, I think the concept is doing more harm than good in our community. The term is easily abused and its meaning has expanded too much. I recommend we permanently taboo “Outside view,” i.e. stop using the word and use more precise, less confused concepts instead. This post explains why.
What does “Outside view” mean now?
Over the past two years I’ve noticed people (including myself!) do lots of different things in the name of the Outside View. I’ve compiled the following lists based on fuzzy memory of hundreds of conversations with dozens of people:
Big List O’ Things People Describe As Outside View:
Reference class forecasting, the practice of computing a probability of an event by looking at the frequency with which similar events occurred in similar situations. Also called comparison class forecasting. [EDIT: Eliezer rightly points out that sometimes reasoning by analogy is undeservedly called reference class forecasting; reference classes are supposed to be held to a much higher standard, in which your sample size is larger and the analogy is especially tight.]
Trend extrapolation, e.g. “AGI implies insane GWP growth; let’s forecast AGI timelines by extrapolating GWP trends.”
Foxy aggregation, the practice of using multiple methods to compute an answer and then making your final forecast be some intuition-weighted average of those methods.
Bias correction, in others or in oneself, e.g. “There’s a selection effect in our community for people who think AI is a big deal, and one reason to think AI is a big deal is if you have short timelines, so I’m going to bump my timelines estimate longer to correct for this.”
Deference to wisdom of the many, e.g. expert surveys, or appeals to the efficient market hypothesis, or to conventional wisdom in some fairly large group of people such as the EA community or Western academia.
Anti-weirdness heuristic, e.g. “How sure are we about all this AI stuff? It’s pretty wild, it sounds like science fiction or doomsday cult material.”
Priors, e.g. “This sort of thing seems like a really rare, surprising sort of event; I guess I’m saying the prior is low / the outside view says it’s unlikely.” Note that I’ve heard this said even in cases where the prior is not generated by a reference class, but rather from raw intuition.
Ajeya’s timelines model (transcript of interview, link to model)
… and probably many more I don’t remember
Big List O’ Things People Describe As Inside View:
Having a gears-level model, e.g. “Language data contains enough structure to learn human-level general intelligence with the right architecture and training setup; GPT-3 + recent theory papers indicate that this should be possible with X more data and compute…”
Having any model at all, e.g. “I model AI progress as a function of compute and clock time, with the probability distribution over how much compute is needed shifting 2 OOMs lower each decade…”
Deference to wisdom of the few, e.g. “the people I trust most on this matter seem to think…”
Intuition-based-on-detailed-imagining, e.g. “When I imagine scaling up current AI architectures by 12 OOMs, I can see them continuing to get better at various tasks but they still wouldn’t be capable of taking over the world.”
Trend extrapolation combined with an argument for why that particular trend is the one to extrapolate, e.g. “Your timelines rely on extrapolating compute trends, but I don’t share your inside view that compute is the main driver of AI progress.”
Drawing on subject matter expertise, e.g. “my inside view, based on my experience in computational neuroscience, is that we are only a decade away from being able to replicate the core principles of the brain.”
Ajeya’s timelines model (Yes, this is on both lists!)
… and probably many more I don’t remember
What did “Outside view” mean originally?
As far as I can tell, it basically meant reference class forecasting. Kaj Sotala tells me the original source of the concept (cited by the Overcoming Bias post that brought it to our community) was this paper. Relevant quote: “The outside view is … essentially ignores the details of the case at hand, and involves no attempt at detailed forecasting of the future history of the project. Instead, it focuses on the statistics of a class of cases chosen to be similar in relevant respects to the present one.” If you look at the text of Superforecasting, the “it basically means reference class forecasting” interpretation holds up. Also, “Outside view” redirects to “reference class forecasting” in Wikipedia.
To head off an anticipated objection: I am not claiming that there is no underlying pattern to the new, expanded meanings of “outside view” and “inside view.” I even have a few ideas about what the pattern is. For example, priors are sometimes based on reference classes, and even when they are instead based on intuition, that too can be thought of as reference class forecasting in the sense that intuition is often just unconscious, fuzzy pattern-matching, and pattern-matching is arguably a sort of reference class forecasting. And Ajeya’s model can be thought of as inside view relative to e.g. GDP extrapolations, while also outside view relative to e.g. deferring to Dario Amodei.
However, it’s easy to see patterns everywhere if you squint. These lists are still pretty diverse. I could print out all the items on both lists and then mix-and-match to create new lists/distinctions, and I bet I could come up with several at least as principled as this one.
This expansion of meaning is bad
When people use “outside view” or “inside view” without clarifying which of the things on the above lists they mean, I am left ignorant of what exactly they are doing and how well-justified it is. People say “On the outside view, X seems unlikely to me.” I then ask them what they mean, and sometimes it turns out they are using some reference class, complete with a dataset. (Example: Tom Davidson’s four reference classes for TAI). Other times it turns out they are just using the anti-weirdness heuristic. Good thing I asked for elaboration!
Separately, various people seem to think that the appropriate way to make forecasts is to (1) use some outside-view methods, (2) use some inside-view methods, but only if you feel like you are an expert in the subject, and then (3) do a weighted sum of them all using your intuition to pick the weights. This is not Tetlock’s advice, nor is it the lesson from the forecasting tournaments, especially if we use the nebulous modern definition of “outside view” instead of the original definition. (For my understanding of his advice and those lessons, see this post, part 5. For an entire book written by Yudkowsky on why the aforementioned forecasting method is bogus, see Inadequate Equilibria, especially this chapter. Also, I wish to emphasize that I myself was one of these people, at least sometimes, up until recently when I noticed what I was doing!)
Finally, I think that too often the good epistemic standing of reference class forecasting is illicitly transferred to the other things in the list above. I already gave the example of the anti-weirdness heuristic; my second example will be bias correction: I sometimes see people go “There’s a bias towards X, so in accordance with the outside view I’m going to bump my estimate away from X.” But this is a different sort of bias correction. To see this, notice how they used intuition to decide how much to bump their estimate, and they didn’t consider other biases towards or away from X. The original lesson was that biases could be corrected by using reference classes. Bias correction via intuition may be a valid technique, but it shouldn’t be called the outside view.
I feel like it’s gotten to the point where, like, only 20% of uses of the term “outside view” involve reference classes. It seems to me that “outside view” has become an applause light and a smokescreen for over-reliance on intuition, the anti-weirdness heuristic, deference to crowd wisdom, correcting for biases in a way that is itself a gateway to more bias…
I considered advocating for a return to the original meaning of “outside view,” i.e. reference class forecasting. But instead I say:
Taboo Outside View; use this list of words instead
I’m not recommending that we stop using reference classes! I love reference classes! I also love trend extrapolation! In fact, for literally every tool on both lists above, I think there are situations where it is appropriate to use that tool. Even the anti-weirdness heuristic.
What I ask is that we stop using the words “outside view” and “inside view.” I encourage everyone to instead be more specific. Here is a big list of more specific words that I’d love to see, along with examples of how to use them:
Reference class forecasting
“I feel like the best reference classes for AGI make it seem pretty far away in expectation.”
“I don’t think there are any good reference classes for AGI, so I think we should use other methods instead.”
Analogy
Analogy is like a reference class but with lower standards; sample size can be small and the similarities can be weaker.
“I’m torn between thinking of AI as a technology vs. as a new intelligent species, but I lean towards the latter.”
Trend extrapolation
“The GWP trend seems pretty relevant and we have good data on it”
“I claim that GPT performance trends are a better guide to AI timelines than compute or GWP or anything else, because they are more directly related.”
Foxy aggregation (a.k.a. multiple methods)
“OK that model is pretty compelling, but to stay foxy I’m only assigning it 50% weight.”
Bias correction
“I feel like things generally take longer than people expect, so I’m going to bump my timelines estimate to correct for this. How much? Eh, 2x longer seems good enough for now, but I really should look for data on this.”
Deference
“I’m deferring to the markets on this one.”
“I think we should defer to the people building AI.”
Anti-weirdness heuristic
“How sure are we about all this AI stuff? The anti-weirdness heuristic is screaming at me here.”
Priors
“This just seems pretty implausible to me, on priors.”
(Ideally, say whether your prior comes from intuition or a reference class or a model. Jia points out “on priors” has similar problems as “on the outside view.”)
i.e. what your view would be if you weren’t deferring to anyone.
“My independent impression is that AGI is super far away, but a lot of people I respect disagree.”
“It seems to me that…”
i.e. what your view would be if you weren’t deferring to anyone or trying to correct for your own biases.
“It seems to me that AGI is just around the corner, but I know I’m probably getting caught up in the hype.”
Alternatively: “I feel like…”
Feel free to end the sentence with “...but I am not super confident” or “...but I may be wrong.”
Subject matter expertise
“My experience with X suggests…”
Models
“The best model, IMO, suggests that…” and “My model is…”
(Though beware, I sometimes hear people say “my model is...” when all they really mean is “I think…”)
Wild guess (a.k.a. Ass-number)
“When I said 50%, that was just a wild guess, I’d probably have said something different if you asked me yesterday.”
Intuition
“It’s not just an ass-number, it’s an intuition! Lol. But seriously though I have thought a lot about this and my intuition seems stable.”
Conclusion
Whenever you notice yourself saying “outside view” or “inside view,” imagine a tiny Daniel Kokotajlo hopping up and down on your shoulder chirping “Taboo outside view.”
Many thanks to the many people who gave comments on a draft: Vojta, Jia, Anthony, Max, Kaj, Steve, and Mark. Also thanks to various people I ran the ideas by earlier.
- The basic reasons I expect AGI ruin by Apr 18, 2023, 3:37 AM; 187 points) (
- CLR’s Annual Report 2021 by Feb 26, 2022, 12:47 PM; 79 points) (EA Forum;
- Prizes for the 2021 Review by Feb 10, 2023, 7:47 PM; 69 points) (
- Resources that (I think) new alignment researchers should know about by Oct 28, 2022, 10:13 PM; 69 points) (
- Voting Results for the 2021 Review by Feb 1, 2023, 8:02 AM; 66 points) (
- Grokking “Semi-informative priors over AI timelines” by Jun 12, 2022, 10:15 PM; 60 points) (EA Forum;
- The basic reasons I expect AGI ruin by Apr 18, 2023, 3:37 AM; 58 points) (EA Forum;
- Winning isn’t enough by Nov 5, 2024, 11:37 AM; 38 points) (
- Some Intuitions Around Short AI Timelines Based on Recent Progress by Apr 11, 2023, 4:23 AM; 37 points) (
- Winning isn’t enough by Nov 5, 2024, 11:43 AM; 31 points) (EA Forum;
- Forecasting Newsletter: June 2021 by Jul 1, 2021, 8:59 PM; 29 points) (EA Forum;
- Resources that (I think) new alignment researchers should know about by Oct 28, 2022, 10:13 PM; 20 points) (EA Forum;
- Some “meta-cruxes” for AI x-risk debates by May 19, 2024, 12:21 AM; 20 points) (
- Mar 3, 2023, 7:33 PM; 17 points) 's comment on Robin Hanson’s latest AI risk position statement by (
- Grokking “Semi-informative priors over AI timelines” by Jun 12, 2022, 10:17 PM; 15 points) (
- What role should evolutionary analogies play in understanding AI takeoff speeds? by Dec 11, 2021, 1:19 AM; 14 points) (
- Forecasting Newsletter: June 2021 by Jul 1, 2021, 9:35 PM; 13 points) (
- Jul 20, 2021, 8:30 AM; 13 points) 's comment on Some thoughts on David Roodman’s GWP model and its relation to AI timelines by (
- What role should evolutionary analogies play in understanding AI takeoff speeds? by Dec 11, 2021, 1:16 AM; 12 points) (EA Forum;
- Dec 6, 2021, 10:31 AM; 11 points) 's comment on Interpreting Yudkowsky on Deep vs Shallow Knowledge by (
- Aug 7, 2021, 10:32 AM; 6 points) 's comment on What are some beautiful, rationalist sounds? by (
- Dec 13, 2021, 10:15 PM; 5 points) 's comment on Evidence on good forecasting practices from the Good Judgment Project: an accompanying blog post by (EA Forum;
- Aug 12, 2021, 1:07 PM; 4 points) 's comment on Automating Auditing: An ambitious concrete technical research proposal by (
- Jan 16, 2024, 11:22 AM; 2 points) 's comment on If Clarity Seems Like Death to Them by (
- Feb 28, 2022, 12:01 AM; 1 point) 's comment on How I Formed My Own Views About AI Safety by (EA Forum;
- Jan 16, 2024, 3:09 AM; 1 point) 's comment on If Clarity Seems Like Death to Them by (
This is a negative review of an admittedly highly-rated post.
The positives first; I think this post is highly reasonable and well written. I’m glad that it exists and think it contributes to the intellectual conversation in rationality. The examples help the reader reason better, and it contains many pieces of advice that I endorse.
But overall, 1) I ultimately disagree with its main point, and 2) it’s way too strong/absolutist about it.
Throughout my life of attempting to have true beliefs and take effective actions, I have quite strongly learned some distinction that maps onto the ideas of inside and outside view. I find this distinction extremely helpful, and specifically, remembering to use (what I call) the outside view often wins me a lot of Bayes points.
When I read through the Big Lists O’ Things, I have these responses;
I think many of those things are simply valid uses of the terms[1]
People using a term wrong isn’t a great reason[2] to taboo that term; e.g. there are countless mis-uses of the concept of “truth” or “entropy” or “capitalism”, but the concepts still carve reality
Seems like maybe some of these you heard one person use once, and then it got to go on the list?
A key example of the absolutism comes from the intro: “I recommend we permanently taboo “Outside view,” i.e. stop using the word and use more precise, less confused concepts instead.” (emphasis added).
But, as described in the original linked sequence post, the purpose of tabooing a word is to remember why you formed a concept in the first place, and see if that break-down helps you reason further. The point is not to stop using a word.
I think the absolutism has caused this post to have negative effects; the phrase “taboo the outside view” has stuck around as a meme, and in my memory, when people use it it has not tended to be good for the conversation.
Instead, I think the post should have said the following.
The term “outside view” can mean many things that can be importantly different (and sometimes wrong)
When you reach for this term, take a second to consider more specifically what you mean, and considering saying that more specific thing instead
When you hear someone say “outside view”, consider asking them more specifically what they mean
(And to be clear, I’m against the absolutism here because I think it is incorrect, not because e.g. rationalists should generally be more polite and hedged, or anything like that.)
I don’t think it makes sense to use “outside view” as a synonym for “reference class forecasting”. The joint that it carves in reality is bigger than that. And I think it’s okay if terms move away from their initial historical use.
To be fair I do actually consider it Bayesian evidence that we should taboo terms; the point of language is to communicate, and if a term is sufficiently mis-used, then it is a net-negative for communication.
Thanks for this thoughtful review! I think the more moderate version that you recommend is very sensible, even though I still stand by the more aggressive original version.
I think a crux for me is just how these things go in practise. Some people have reported that the meme “taboo outside view” has been helpful in conversations, and now you are reporting that it was harmful. I’m interested to hear that & would be interested to hear more details/examples if you remember any. I suppose ideally we’d do a poll of the community to find out whether the taboo has been overzealously applied or not.
I’m also interested in examples (I actually haven’t seen the phrase “taboo outside view” in the wild, one way or another)
What considerations might lead you to not say the more specific thing? Can you give a few examples of cases where it’s better to say “outside view” than to say something more specific?
If “outside view” was a natural category that was useful to use, AND people generally had a shared understanding of what it meant, then it would slow things down unnecessarily to be more specific, at least often (sometimes even then you’d want to be more specific.) My original post cast doubt on both the naturalness/usefulness of the concept (not saying there’s absolutely nothing tying the things in the Big Lists together, just saying that there isn’t really any good evidence that whatever it is that ties them together is epistemically important) and the shared understanding (ho boy do different people seem to have different ideas of what it means and how it should be used and what evidential status it confers)
Yeah, I agree with all of this; see my own review. My guess is that Alex_Altair is making the exact mistake you tried to warn against. But, if I’m wrong, the examples would have been clarifying.
This post is worthwhile and correct, with clear downstream impact. It might be the only non-AI post of 2021 that I’ve heard cited in in-person conversation—and the cite immediately improved the discussion.
It’s clearly written and laid out; unless you’re already an excellent technical writer, you can probably learn something by ignoring its content and studying its structure.
I liked this post when I read it. It matched my sense that (e.g.) using “outside view” to refer to Hanson’s phase transition model of agriculture->industry->AI was overstating the strength of the reasoning behind it.
But I’ve found that I’ve continued to use the terms “inside view” and “outside view” to refer to the broad categories sketched out in the two Big Lists O’ Things. Both in my head and when speaking. (Or I’ll use variants like “outside viewish” or similar.)
I think there is a meaningful distinction here: the reasoning moves on the “Outside” list involve looking at patterns, getting a sense of the general shape of things, thinking evidentially; the reasoning moves on the “Inside” list involve looking at mechanisms, relying on very specific claims/pictures of how things work, thinking causally. “Outside” and “Inside” actually seem like pretty apt labels for this distinction. And tracking this distinction can be helpful for keeping track of what an argument is and isn’t trying to do.
For me, the distinction doesn’t carry the implication that one is good and the other is not, or that the “outside” category is Backed By Science, or that an outside-style line of thinking represents the one Outside View on a question, or that categorizing an argument as inside or outside is all that precise a characterization of what it’s based on.