Call For Distillers
Many technical alignment researchers are bad-to-mediocre at writing up their ideas and results in a form intelligible to other people. And even for those who are reasonably good at it, writing up a good intuitive explanation still takes a lot of work, and that work lengthens the turn-time on publishing new results. For instance, a couple months ago I wrote a post which formalized the idea of abstractions as redundant information, and argued that it’s equivalent to abstractions as information relevant at a distance. That post came out about two months after I had the rough math worked out, because it took a lot of work to explain it decently—and I don’t even think the end result was all that good an explanation! And I still don’t have a post which explains well why that result is interesting.
I think there’s a lot of potential space in the field for people who are good at figuring out what other researchers’ math is saying intuitively, and why it’s interesting, and then communicating that clearly—i.e. the skill of distillation. This post will briefly sketch out what two kinds of distillation roles might look like, what skills are needed, and talk about how one might get started in such a role.
Two Distiller Roles
The two types of distiller role I’ll sketch are:
“Independent” distiller: someone who works independently, understanding work published by other researchers and producing distillations of that work.
“Adjunct” distiller: someone who works directly with one researcher or a small team, producing regular write-ups of what the person/team is thinking about and why.
These two roles add value in slightly different ways.
An independent distiller’s main value-adds are:
Explaining the motivation and intended applications
Coming up with new examples
Boiling down the “key intuitive story” behind an argument
Showing how the intuitive story fits into the context of the intended applications
I expect the ability to come up with novel examples and boil down the core intuitive story behind a bunch of math are the rate-limiting skills here.
Rob Miles is a good example of an existing independent distiller in the field. He makes YouTube videos intuitively explaining various technical results and arguments. Rob’s work is aimed somewhat more at a popular audience than what I have in mind, but it’s nonetheless been useful for people in the field.
I expect an adjunct distiller’s main value-adds are:
Writing up explanations, examples, and intuitions, similar to the independent distiller
Saving time for the technical researcher/team; allow more specialization
Providing more external visibility/legibility into the research process and motivation
Accelerating the research process directly by coming up with good examples and intuitive explanations
I expect finding a researcher/team to work with is the rate-limiting step to this sort of work.
Mark Xu is a good example of an existing adjunct distiller. He’s worked with both Evan Hubinger and Paul Christiano, and has written up decent distillations of some of their thoughts. I believe Mark did this with the aim of later doing technical research himself, rather than mostly being a distiller. That is a pretty good strategy and I expect it to be a common pathway, though naturally I expect people who aim to specialize in distillation long-term will end up better at distillation.
What Kind Of Skills Are Needed?
I expect the key rate-limiting skills are:
Ability to independently generate intuitive examples when reading mathematical arguments, or having a mathematical discussion
Ability to extract the core intuitive story from a mathematical argument
Writing/drawing skills to clearly convey technical intuitions to a wider audience
Ability to do most of the work of crossing the communication gap yourself—both so that researchers do not need to spend a lot of effort communicating to you, and so that readers do not need to spend a lot of effort understanding you
For the adjunct role, ability to write decent things quickly and frequently without too much perfectionism
For the non-adjunct role, ability to do all this relatively independently
How To Get Started
Getting started in an independent distiller role should be pretty straightforward: choose some research, and produce some distillations. It’s inherently a very legible job, so you should pretty quickly have some good example pieces which you could showcase in a grant application (e.g. from the Long Term Future Fund or FTX Future Fund). That said, bear in mind that you may need some practice before you actually start to produce very good distillations.
An adjunct role is more difficult, because you need someone to work with. Obvious advice: just asking people is an underutilized strategy, and works surprisingly well. Be sure to emphasize your intended value-add to the researcher(s). If you want to prove yourself a bit before reaching out, independently distilling some of a researcher’s existing public work is another obvious step. You might also try interviewing a researcher on some part of their work, and then distilling that, in order to get a better feel for what it would be like to work together before actually committing.
- How to pursue a career in technical AI alignment by 4 Jun 2022 21:36 UTC; 265 points) (EA Forum;
- Finite Factored Sets in Pictures by 11 Dec 2022 18:49 UTC; 174 points) (
- Reflections on my 5-month AI alignment upskilling grant by 28 Dec 2022 7:23 UTC; 113 points) (EA Forum;
- Calling for Student Submissions: AI Safety Distillation Contest by 23 Apr 2022 20:24 UTC; 102 points) (EA Forum;
- Reflections on my 5-month alignment upskilling grant by 27 Dec 2022 10:51 UTC; 82 points) (
- Call For Distillers by 6 Apr 2022 3:03 UTC; 70 points) (EA Forum;
- How to pursue a career in technical AI alignment by 4 Jun 2022 21:11 UTC; 69 points) (
- Another call for EA distillers by 3 Aug 2022 14:29 UTC; 68 points) (EA Forum;
- A tough career decision by 9 Apr 2022 0:46 UTC; 68 points) (EA Forum;
- “Brain enthusiasts” in AI Safety by 18 Jun 2022 9:59 UTC; 63 points) (
- Voting Results for the 2022 Review by 2 Feb 2024 20:34 UTC; 57 points) (
- AI safety university groups: a promising opportunity to reduce existential risk by 30 Jun 2022 18:37 UTC; 53 points) (EA Forum;
- 2022 (and All Time) Posts by Pingback Count by 16 Dec 2023 21:17 UTC; 53 points) (
- Calling for Student Submissions: AI Safety Distillation Contest by 24 Apr 2022 1:53 UTC; 48 points) (
- Distillation Contest—Results and Recap by 26 Jul 2022 20:30 UTC; 47 points) (EA Forum;
- A multi-disciplinary view on AI safety research by 8 Feb 2023 16:50 UTC; 43 points) (
- List of technical AI safety exercises and projects by 19 Jan 2023 9:35 UTC; 41 points) (
- Distillation Contest—Results and Recap by 29 Jul 2022 17:40 UTC; 34 points) (
- 29 Apr 2022 4:21 UTC; 31 points) 's comment on Increasing Demandingness in EA by (EA Forum;
- What I learned at the AI Safety Europe Retreat by 17 Apr 2023 17:40 UTC; 28 points) (
- How to get people to produce more great exposition? Some strategies and their assumptions by 25 May 2022 22:30 UTC; 26 points) (
- General advice for transitioning into Theoretical AI Safety by 15 Sep 2022 5:23 UTC; 25 points) (EA Forum;
- 11 May 2022 20:44 UTC; 23 points) 's comment on What is your theory of victory? by (EA Forum;
- Announcing the Distillation for Alignment Practicum (DAP) by 18 Aug 2022 19:50 UTC; 23 points) (
- [Request for Distillation] Coherence of Distributed Decisions With Different Inputs Implies Conditioning by 25 Apr 2022 17:01 UTC; 22 points) (
- AI Risk Intro 2: Solving The Problem by 22 Sep 2022 13:55 UTC; 22 points) (
- Announcing the Distillation for Alignment Practicum (DAP) by 18 Aug 2022 22:31 UTC; 20 points) (EA Forum;
- Disagreements about Alignment: Why, and how, we should try to solve them by 8 Aug 2022 22:32 UTC; 16 points) (EA Forum;
- List of technical AI safety exercises and projects by 19 Jan 2023 9:35 UTC; 15 points) (EA Forum;
- A distillation of Evan Hubinger’s training stories (for SERI MATS) by 18 Jul 2022 3:38 UTC; 15 points) (
- Request for beta readers by 6 Apr 2022 5:29 UTC; 14 points) (
- AI safety university groups: a promising opportunity to reduce existential risk by 1 Jul 2022 3:59 UTC; 14 points) (
- General advice for transitioning into Theoretical AI Safety by 15 Sep 2022 5:23 UTC; 12 points) (
- AI Risk Intro 2: Solving The Problem by 24 Sep 2022 9:33 UTC; 11 points) (EA Forum;
- Disagreements about Alignment: Why, and how, we should try to solve them by 9 Aug 2022 18:49 UTC; 11 points) (
- 10 Nov 2022 15:38 UTC; 5 points) 's comment on Some advice on independent research by (
- Which one of these two academic routes should I take to end up in AI Safety? by 3 Jul 2022 1:05 UTC; 5 points) (
- 26 Apr 2022 11:26 UTC; 4 points) 's comment on Intuitions about solving hard problems by (
- 2 Jan 2023 7:49 UTC; 3 points) 's comment on Alignment, Anger, and Love: Preparing for the Emergence of Superintelligent AI by (
- 8 Apr 2022 10:35 UTC; 2 points) 's comment on How to write a LW sequence to learn a topic? by (
Since this post was written, I feel like there’s been a zeitgeist of “Distillation Projects.” I don’t know how causal this post was, I think in some sense the ecosystem was ripe for a Distillation Wave) But it seemed useful to think about how that wave played out.
Some of the results have been great. But many of the results have felt kinda meh to me, and I now have a bit of a flinch/ugh reaction when I see a post with “distillation” in it’s title.
Basically, good distillations are a highly skilled effort. It’s sort of natural to write a distillation of something as part of your attempt to understand it, and upskill (I think I had previously advocated this sometimes). I think this was basically a reasonable thing to do, but, it did have the cumulative effect of decreasing the signal-noise ratio of distillations, since most people doing this aren’t skilled yet.
None of that contradicts the claims of this post, which specify various skills you need, and recommends actually investing in those skills. (The title of this post is “call for Distillers”, not “call for Distillations”. I think a lot of what happened was things like “Distillation contests”, and incorporating distillation into SERI MATS programming, etc, which doesn’t automatically produce dedicated distillers)