DeepMind is hiring for the Scalable Alignment and Alignment Teams
We are hiring for several roles in the Scalable Alignment and Alignment Teams at DeepMind, two of the subteams of DeepMind Technical AGI Safety trying to make artificial general intelligence go well. In brief,
The Alignment Team investigates how to avoid failures of intent alignment, operationalized as a situation in which an AI system knowingly acts against the wishes of its designers. Alignment is hiring for Research Scientist and Research Engineer positions.
The Scalable Alignment Team (SAT) works to make highly capable agents do what humans want, even when it is difficult for humans to know what that is. This means we want to remove subtle biases, factual errors, or deceptive behaviour even if they would normally go unnoticed by humans, whether due to reasoning failures or biases in humans or due to very capable behaviour by the agents. SAT is hiring for Research Scientist—Machine Learning, Research Scientist—Cognitive Science, Research Engineer, and Software Engineer positions.
We elaborate on the problem breakdown between Alignment and Scalable Alignment next, and discuss details of the various positions.
“Alignment” vs “Scalable Alignment”
Very roughly, the split between Alignment and Scalable Alignment reflects the following decomposition:
Generate approaches to AI alignment – Alignment Team
Make those approaches scale – Scalable Alignment Team
In practice, this means the Alignment Team has many small projects going on simultaneously, reflecting a portfolio-based approach, while the Scalable Alignment Team has fewer, more focused projects aimed at scaling the most promising approaches to the strongest models available.
Scalable Alignment’s current approach: make AI critique itself
Imagine a default approach to building AI agents that do what humans want:
Pretrain on a task like “predict text from the internet”, producing a highly capable model such as Chinchilla or Flamingo.
Fine-tune into an agent that does useful tasks, as evaluated by human judgements.
There are several ways this could go wrong:
Humans are unreliable: The human judgements we train against could be flawed: we could miss subtle factual errors, use biased reasoning, or have insufficient context to evaluate the task.
The agent’s reasoning could be hidden: We want to know not just what the system is doing but why, both because that might reveal something about what that we don’t like, and because we expect good reasoning to better generalize to other situations.
Even if the agent is reasoning well, it could fail in other situations: Even if the reasoning is correct this time, the AI could fail to generalize correctly to other situations.
Our current plan to address these problem is (in part):
Give humans help in supervising strong agents: On the human side, provide channels for oversight and advice from peers, experts in various domains, and broader society. On the ML side, agents should explain their behaviour and reasoning, argue against themselves when wrong, and cite relevant evidence.
Align explanations with the true reasoning process of the agent: Ensure that agent’s are able and incentivized to show their reasoning to human supervisors, either by making reasoning explicit if possible or via methods for interpretability and eliciting latent knowledge.
Red team models to exhibit failure modes that don’t occur in normal use
We believe none of these pieces are sufficient by themselves:
(1) without (2) can be rationalization, where an agent decides what to do and produces an explanation after the fact that justifies its answer.
(2) without (1) doesn’t scale: The full reasoning trace of the agent might be enormous, it might be terabytes of data even with compression, or exponentially large without compression if the agent is using advanced heuristics which expand into very large human-interpretable reasoning traces.
(1)+(2) without (3) will miss rare failures.
(3) needs (1)+(2) to define failure.
An example proposal for (1) is debate, in which two agents are trained in a zero-sum game to provide evidence and counterarguments for answers, as evaluated by a human judge. If we imagine the exponentially large tree of all possible debates, the goals of debate are to (1) engineer the whole tree so that it captures all relevant considerations and (2) train agents so that the chosen single path through the tree reflects the tree as a whole.
The full picture will differ from the pure debate setting in many ways, and we believe the correct interpretation of the debate idea is “agents should critique themselves”. There is a large space of protocols that include agents critiquing agents as a component, and choosing between them will involve
Human aspects: Whether a particular human-machine interaction is aligned depends on the humans involved, and we need cognitive science experiments probing these uncertainties.
Theoretical strengthenings: There are a variety of theoretical strengthenings to debate such as cross-examination, learning the prior, and market making, but little work has been done to turn these into practical systems (example obstacle to learning the prior).
Flexibility in integrating other components of alignment: For example, if strong interpretability tools are developed they should be smoothly integrated into the human-machine interaction, so that the human supervision process has access to internal reasoning.
Practicalities: Citing sources, better uncertainty estimation, declining to answer if uncertain, etc.
Governance: By explaining themselves, agents can provide a lever for external oversight.
The three goals of “help humans with supervision”, “align explanations with reasoning”, and “red teams” will be blurry once we put the whole picture together. Red teaming can occur either standalone or as an integrated part of a training scheme such as cross-examination, which allows agents to interrogate opponent behavior along counterfactual trajectories. Stronger schemes to help humans with supervision should improve alignment with reasoning by themselves, as they grow the space of considerations that can be exposed to humans. Thus, a key part of the Scalable Alignment Team’s work is planning out how these pieces will fit together.
Examples of our work, involving extensive collaboration with other teams at DeepMind:
Risk analyses, both for long-term alignment risks and harms that exist today:
Language model pretraining, analysis, and safety discussion
Safety
Earlier proposals for debate and human aspects of debate
We view our recent safety papers as steps towards the broader scalable alignment picture, and continue to build out towards debate and generalizations. We work primarily with large language models (LLMs), both because LLMs are a tool for safety by enabling human-machine communication and are examples of ML models that may cause both near-term and long-term harms.
Alignment Team’s portfolio of projects
In contrast to the Scalable Alignment Team, the Alignment Team explores a wide variety of possible angles on the AI alignment problem. Relative to Scalable Alignment, we check whether a technique could plausibly scale based on conceptual and abstract arguments. This lets us iterate much faster at the cost of getting less useful feedback from reality. To give you a sense of the variety, here’s some examples of public past work that was led by current team members:
Learning objectives from human feedback on hypothetical behavior
Understanding agent incentives using causal influence diagrams
Improving our philosophical understanding of “agency” using Conway’s game of life
That being said, over the last year there has been some movement away from previous research topics and towards others. To get a sense of our current priorities, here are short descriptions of some projects that we are currently working on:
Primarily conceptual:
Investigate threat models in which due to increasing AI sophistication, humans are forced to rely on evaluations of outcomes (rather than evaluations of process or reasoning).
Investigate arguments about the difficulty of AI alignment, including as a subproblem the likelihood that various AI alignment plans succeed.
Compare various decompositions of the alignment problem to see which one is most useful for guiding future work.
Primarily empirical:
Create demonstrations of inner alignment failures, in a similar style as this paper.
Dig deeper into the grokking phenomenon and give a satisfying account of how and why it happens.
Develop interpretability tools that allow us to understand how large language models work (along similar lines as Anthropic’s work).
Evaluate how useful process-based feedback is on an existing benchmark.
Relative to most other teams at DeepMind, on the Alignment team there is quite a lot of freedom in what you work on. All you need to do to start a project is to convince your manager that it’s worth doing (i.e. reduces x-risk comparably well to other actions you could take), and convince enough collaborators to work on the project.
In many ways the team is a collection of people with very different research agendas and perspectives on AI alignment that you wouldn’t normally expect to work together. What ties us together is our meta-level focus on reducing existential risk through alignment failures:
Every new project must come accompanied by a theory of change that explains how it reduces existential risk; this helps us avoid the failure mode of working on interesting conceptual projects that end up not connecting to the situations we are worried about.
It’s encouraged to talk to people on the team with very different perspectives and try to come to agreement, or at least better understand each other’s positions. This can be an explicit project even though it isn’t “research” in the traditional sense.
Interfacing with the rest of DeepMind
Both Alignment and Scalable Alignment collaborate extensively with people across DeepMind.
For Alignment, this includes both collaborating on projects that we think are useful, and by explaining our ideas to other researchers. As a particularly good example, we recently ran a 2 hour AI alignment “workshop” with over 100 attendees. (That being said, you can opt out of these engagements in order to focus on research, if you prefer.)
As Scalable Alignment’s work with large language models is very concrete, we have tight collaborations with a variety of teams, including large-scale pretraining and other language teams, Ethics and Society, and Strategy and Governance.
The roles
Between our two teams we have open roles for Research Scientists (RSs), Research Engineers (REs), and (for Scalable Alignment) Software Engineers. Scalable Alignment RSs can have either a machine learning background or a cognitive science background (or equivalent). The boundaries between these roles are blurry. There are many skills involved in overall Alignment / Scalable Alignment research success: proposing and leading projects, writing and publishing papers, conceptual safety work, algorithm design and implementation, experiment execution and tuning, design and implementation of flexible, high-performance, maintainable software, and design and analysis of human interaction experiments.
We want to hire from the Pareto frontier of all relevant skills. This means RSs are expected to have more research experience and more of a track record of papers, but SWEs are expected to be better at scalable software design / collaboration / implementation, with REs in between, but also that REs can and do propose and lead projects if capable (e.g., this recent paper had an RE as last author). For more details on the tradeoffs, see the career section of Rohin’s FAQ.
For Scalable Alignment, most of our work focuses on large language models. For Machine Learning RSs, this means experience with natural language processing is valuable, but not required. We are also interested in candidates motivated by other types of harms caused by large models, such as those described in Weidinger et al., Ethical and social risks of harm from language models, as long as you are excited by the goal of removing such harms even in subtle cases which humans have difficulty detecting. For REs and SWEs, a focus on large language models means that experience with high performance computation or large, many-developer codebases is valuable. For the RE role for Alignment, many of the projects you could work on would involve smaller models that are less of an engineering challenge, though there are still a few projects that work with our largest language models.
Scalable Alignment Cognitive Scientists are expected to have a track record of research in cognitive science, and to design, lead, and implement either standalone human-only experiments to probe uncertainty, or the human interaction components of mixed human / machine experiments. No experience with machine learning is required, but you should be excited to collaborate with people who do!
Apply now!
We will be evaluating applications on a rolling basis until positions are filled, but we will at least consider all applications that we receive by May 31. Please do apply even if your start date is up to a year in the future, as we probably will not run another hiring round this year. These roles are based in London, with a hybrid work-from-office / work-from-home model. International applications are welcome as long as you are willing to relocate to London.
While we do expect these roles to be competitive, we have found that people often overestimate how much we are looking for. In particular:
We do not expect you to have a PhD if you are applying for the Research Engineer or Software Engineer roles. Even for the Research Scientist role, it is fine if you don’t have a PhD if you can demonstrate comparable research skill (though we do not expect to see such candidates in practice).
We do not expect you to have read hundreds of blog posts and papers about AI alignment, or to have a research agenda that aims to fully solve AI alignment. We will look for understanding of the basic motivation for AI alignment, and the ability to reason conceptually about future AI systems that we haven’t yet built.
If we ask you, say, whether an assistive agent would gradient hack if it learned about its own training process, we’re looking to see how you go about thinking about a confusing and ill-specified question (which happens all the time in alignment research). We aren’t expecting you to give us the Correct Answer, and in fact there isn’t a correct answer; the question isn’t specified well enough for that. We aren’t even expecting you to know all the terms; it would be fine to ask what we mean by “gradient hacking”.
As a rough test for the Research Engineer role, if you can reproduce a typical ML paper in a few hundred hours and your interests align with ours, we’re probably interested in interviewing you.
We do not expect SWE candidates to have experience with ML, but you should have experience with high performance code and experience with large, collaborative codebases (including the human aspects of collaborative software projects).
Go forth and apply!
Alignment Team:
Scalable Alignment Team:
- Shallow review of live agendas in alignment & safety by 27 Nov 2023 11:10 UTC; 332 points) (
- How to pursue a career in technical AI alignment by 4 Jun 2022 21:36 UTC; 265 points) (EA Forum;
- AGI Safety and Alignment at Google DeepMind: A Summary of Recent Work by 20 Aug 2024 16:22 UTC; 221 points) (
- Shallow review of live agendas in alignment & safety by 27 Nov 2023 11:33 UTC; 76 points) (EA Forum;
- How to pursue a career in technical AI alignment by 4 Jun 2022 21:11 UTC; 69 points) (
- The Importance of AI Alignment, explained in 5 points by 11 Feb 2023 2:56 UTC; 50 points) (EA Forum;
- AI safety technical research—Career review by 17 Jul 2023 15:34 UTC; 49 points) (EA Forum;
- [AN #173] Recent language model results from DeepMind by 21 Jul 2022 2:30 UTC; 37 points) (
- The Importance of AI Alignment, explained in 5 points by 11 Feb 2023 2:56 UTC; 33 points) (
- AI safety technical research—Career review by 17 Jul 2023 15:34 UTC; 14 points) (
- 14 May 2023 23:32 UTC; 11 points) 's comment on How to pursue a career in technical AI alignment by (EA Forum;
- 2 Sep 2022 5:42 UTC; 7 points) 's comment on Levelling Up in AI Safety Research Engineering by (EA Forum;
- Lab Collaboration on AI Safety Best Practices by 17 Mar 2024 12:20 UTC; 3 points) (EA Forum;
Good timing! First scare the hell out of everyone, then use it to hire employees the next day. I like it. (no sarcasm)
Is there any option for remote work? I’m sure there’s plenty of people in less wrong who would live to work on alignment at deep mind, but don’t happen to live in London...
Unfortunately not, though as Tom L points out below, if your concern is about getting a visa, that’s relatively easy to do. DeepMind will provide assistance with the process. I went through it myself and it was relatively painless; it probably took 5-10 hours of my time total (including e.g. travel to and from the appointment where they collected biometric data).
Whilst that’s definitely great, my guess is that 90% of the people who would be interested and don’t live in London, would not move to London for the job, even with a free Visa. Not supporting remote work therefore loses out on a majority of the potential talent pool for this job.
I don’t have a strong opinion on whether it is good to support remote work. I agree we lose out on a lot of potential talent, but we also gain productivity benefits from in person collaboration.
However, this is a DeepMind-wide policy and I’m definitely not sold enough on the importance of supporting remote work to try and push for an exception here.
Pretty disappointing and unexpected to hear this in 2022, after all the learnings from the pandemic.
I can’t speak to the option for remote work but as a counterpoint, it seems very straightforward to get a UK visa for you and your spouse/children (at least straightforward relative to the US). The relevant visa to google is the Skilled Worker / Tier 2 visa if you want to know more.
ETA: Of course, there are still legitimate reasons for not wanting to move. Just wanted to point out that the legal barrier is lower than you might think.
I’m not familiar with London housing prices. Is it possible to affordably rent or mortgage a decent 2 bedroom condo within a 5 minute walk of the offices with your compensation package? (By affordable I mean less than 1⁄3 of total comp spent on housing, stretching to 1⁄2 if comp is unusually high.)
Whilst you specified a 5 minute walk on your criteria I think you should also consider the fact that there is generally very good public transport in London where prices would be cheaper a bit further out. The location is close to several rail stations (Blackfriars, Farringdon), the new Crossrail route (farringdon) opening in a few weeks and tube lines (central, circle, district). With the new cross-rail route it should be possible to access zone 4 locations in 20 minutes.
Almost certainly, e.g. this one meets those criteria and I’m pretty sure costs < 1⁄3 of total comp (before taxes), though I don’t actually know what typical total comp is. You would find significantly cheaper places if you were willing to compromise on commute, since DeepMind is right in the center of London.
Thanks, that is more luxurious than I imagined, so families should have no difficulty finding a large enough place.
I’m potentially interested in the Research Engineer position on the Alignment Team, but I’m currently 3 months into a 6 month grant from LTFF to reorient my career from general machine learning to AI safety specifically. My current plan is to keep doing solo work () until the last month of my grant period then begin applying to AI safety work at places like Anthropic, Redwood Research, Open AI, and Deepmind.
Do you think there’s a significant advantage to applying soon vs 3 months from now?
Looking into it, I’ll try to get you a better answer soon. My current best guess is that you should apply 3 months from now. This runs an increased risk that we’ll have filled all our positions / closed our applications, but also improved chances of making it through because you’ll know more things and be better prepared for the interviews.
(Among other things I’m looking into: would it be reasonable to apply now and mention that you’d prefer to be interviewed in 3 months.)
Thanks Rohin. I also feel that interviewing after my 3 more months of independent work is probably the correct call.
Update: I think you should apply now and mention somewhere that you’d prefer to be interviewed in 3 months because in those 3 months you will be doing <whatever it is you’re planning to do> and it will help with interviewing.
The Research Engineer job for the Alignment team is no longer open—is this because it’s reached some threshold of applications? In any case might not be helpful to advertise!
Thanks for doing this though, the context is very useful (I’ve applied as RE to both).
Should be fixed now!
Hi there,
Given that you’ve described various ‘primarily conceptual’ projects on the Alignment Team, and given the distinction between Scientists and Engineers, one aspect that I’m unsure about is roughly: Would you expect a Research Scientist on the Alignment Team to necessarily have a minimum level of practical ML knowledge? Are you able to say any more about that? e.g. Would they have to pass a general Deep Mind coding test or something like that?
Yes, we’d definitely expect some practical ML knowledge, and yes, there would be a coding interview.
There isn’t a set bar; I could imagine hiring someone who did merely okay on the ML interview and was kinda bad on the coding interview, but both their past work and other interviews demonstrated excellent conceptual reasoning skills, precisely because we have many primarily conceptual projects. But I don’t trust conceptual reasoning to get the answer right, if you have very little understanding of how AI systems work in practice, so I still care about some level of ML knowledge.
Thanks Rohin!
Hi. Is the Cognitive Scientist position still open? The LinkedIn ad says “No longer accepting applications”. Thank you!
I believe it’s meant to be open, and this link still seems to work.
For the Cognitive Science Research Scientist position, the application has a separate required sections for explaining (i) why we are interested in joining the alignment team and (ii) our research interests and topics we’d like to explore at DeepMind.
Should the cover letter, which is also required separately, not discuss either of these topics then? Usually, I would think answers to such questions would be part of a cover letter.
Thanks again!
Levan
It’s totally fine to just not provide a cover letter if you think there isn’t anything else you’d like to add to your application.
I’ve found that when we didn’t ask these additional questions, people would often not give a cover letter, or give a cover letter that doesn’t address these topics, and then I find it hard to distinguish (1) the people who didn’t bother explaining their interest in alignment from (2) the people who aren’t interested in alignment. (Given how many applications we get, we need to filter out group (2) before the interviewing stage.) That’s why we ask the questions explicitly now.
Thank you!
Just to confirm, applications received by the 31st (so today) will be considered, but later ones might not, would that be correct?
That’s correct, though now that it is the 31st, I think it’s pretty safe to say that we’ll still consider applications at least up to, say, a week from now (June 7th).
I sent mines in around 11 days ago and haven’t heard back yet. Do you guys have a backlog?
Yes, sorry about that, it’s mostly my fault (I’ve had two back-to-back workshops).
Hi Rohin, I have a question about the technical interviews, stage 2, math/stats, ML/AI, CS (Research Engineer):
Is this stage composed of factual/theoretical questions? Similar to a QUIZ?
Does DeepMind offers a list of materials for preparation? (I’ve received an Alignment_RE_Process.pdf, however, it doesn’t have a guideline for books and courses)
The math/stats and ML/AI are more like factual / theoretical questions, though you might have to derive some results in the interview itself.
Your recruiter should have sent you a standard email with details and advice; I don’t have anything in particular to add (I probably know less than the recruiters for this particular stage of the interview process, since it is standardized across DeepMind).
Hi Rohin, how long does it usually take to hear back if selected for the next stage? I applied two weeks ago but didn’t receive any other mail yet, so I was just curious if I still have a chance or was not selected.
That’s probably my fault, I’ve fallen behind on looking through initial applications. Ping me again if you haven’t gotten a response in another two weeks.
Hi Rohin, I am very interested in the Research Scientist position on the Scalable Alignment Team and have some questions:
Does this process follow the same traditional DeepMind hiring process?
Phase one—initial chat with the recruitment team
Phase two—technical interviews (Technical quiz and Coding interview)
Phase three—research interviews
Phase four—culture interview
If so, can I schedule the first interview in 3 months and use that time to finish current projects and also to prepare? Is there a risk of positions being filled or closed?
(I got a contact from the recruiter yesterday asking me to set my availability for the two 30 minute introductory interviews).
Yes, this follows the traditional DeepMind hiring process. (I think phases three and four happen simultaneously though.)
There is a risk of positions being filled or closed, but it should be fine to ask to defer the interviews until the positions are close to being filled or closed, so that you would still be considered. If you’ve already applied you can let your recruiter know and they should be able to add it as a note to the application.