SERI ML Alignment Theory Scholars Program 2022
The Stanford Existential Risks Initiative (SERI) recently opened applications for the second iteration of the ML Alignment Theory Scholars (MATS) Program, which aims to help aspiring alignment researchers enter the field by pairing them with established research mentors and fostering an academic community in Berkeley, California over the summer. Current mentors include Alex Gray, Beth Barnes, Evan Hubinger, John Wentworth, Leo Gao and Stuart Armstrong. Applications close on May 15 and include a written response to mentor-specific selection questions, viewable on our website.
Who is this program for?
Our ideal applicant has:
an understanding of the AI alignment research landscape equivalent to having chttps://www.agisafetyfundamentals.com/ai-alignment-curriculumompleted the AGI Safety Fundamentals course;
previous experience with technical research (e.g. ML, CS, maths, physics, neuroscience, etc.);
strong motivation to pursue a career in AI alignment research.
For the first stage of the program, we asked each alignment researcher to provide a set of questions that are sufficient to select candidates they would be happy to mentor. Applicants can apply for multiple mentors, but will have to complete each mentor’s selection questions.
What will this program involve?
Over four weeks, the participants will develop an understanding of a research agenda at the forefront of AI alignment through online readings and cohort discussions, averaging 10 h/week from Jun 6 to Jul 1. After this initial upskilling period, the scholars will be paired with an established AI alignment researcher for a two-week “research sprint” to test fit from Jul 4 to Jul 15. Assuming all goes well, scholars will be accepted into an eight-week intensive research program in Berkeley, California over the US summer break (Jul 25 to Sep 16).
Participants will obtain a $6,000 grant for completing the training and research sprint and $16,000 at the conclusion of the program. Furthermore, all expenses will be covered, including accommodation, office space and networking events with the Bay Area alignment community. We are happy to continue providing funding after the two month period to promising scholars, at the discretion of our research mentors. International students can apply to the program, and will arrive in the US under a B1 visa.
We hope to run another iteration of the program in the winter, and possibly in the fall. If you are not able to apply for the summer program, we encourage you to apply for the fall or winter. We may be able to offer different types of visas in future iterations.
Theory of change
This section is intended to explain the reasoning behind our program structure and is not required reading for any applicant. SERI MATS’ theory of change is as follows:
We believe that AI alignment research is pre-paradigmatic, with a diversity of potentially promising research agendas. Therefore, we aim to support many different alignment research agendas to decorrelate failure. We also aim to accelerate the development of scholars into researchers capable of pursuing original agendas and mentoring further scholars.
We believe that working 1:1 with a mentor is the best and quickest way to develop the ability to conduct alignment theory research—that reading curriculum alone is worse for a large number of participants. Moreover, we believe that our target scholars might be able to produce value directly for the mentors by acting as research assistants. For the first few months, we are generally more excited about mentees working on an established mentor’s research agenda than on their own.
We believe that our limiting constraint is mentor time. This means we wish to have strong filtering mechanisms (e.g. candidate selection questions) to ensure that each applicant is suitable for each mentor. We’d rather risk rejecting a strong participant than admitting a weak participant. We offer the possibility for mentors to leave the program at any time they want.
We believe that MATS should be a “mentor-centered” program, in that we are willing to be very flexible regarding mentors’ preferences regarding the structure and implementation of the program.
We believe that there exists a large population of possible alignment researchers whose limitations are not some innate lack of talent, but rather more mundane barriers, which we can address:
Lack of networking within the community to find mentors;
Lack of peers and cohort to discuss research with;
Financial stability; or
Low risk tolerance.
We believe that creating a strong alignment theory community, where scholars share housing and offices, could be extremely beneficial for the development of new ideas. We have already seen promising results of alignment theory collaboration at the office space and housing we provided for the first iteration of SERI MATS and hope to see more!
We are happy to hear any feedback on our aims or strategy. If you would like to become a mentor or join MATS as a program organiser for future program iterations, please send us an email at exec@serimats.org.
- Conditioning Generative Models for Alignment by 18 Jul 2022 7:11 UTC; 59 points) (
- Deception?! I ain’t got time for that! by 18 Jul 2022 0:06 UTC; 55 points) (
- Getting up to Speed on the Speed Prior in 2022 by 28 Dec 2022 7:49 UTC; 36 points) (
- Auditing games for high-level interpretability by 1 Nov 2022 10:44 UTC; 33 points) (
- Training goals for large language models by 18 Jul 2022 7:09 UTC; 28 points) (
- A distillation of Evan Hubinger’s training stories (for SERI MATS) by 18 Jul 2022 3:38 UTC; 15 points) (
- [Job Ad] SERI MATS is (still) hiring for our summer program by 6 Jun 2023 21:07 UTC; 12 points) (
- [Job Ad] MATS is hiring! by 9 Oct 2024 2:17 UTC; 10 points) (
- [Job Ad] SERI MATS is hiring for our summer program by 26 May 2023 4:51 UTC; 8 points) (EA Forum;
I frontpaged this post. I am currently confused as to what our policy should be regarding announcements like this – they are not “timeless” which is the usual frontpage criteria [1], but also I think they are of general interest to a lot of readers. I would feel bad restricting their visibility. In this case, the Theory of Change feels like it’s of more lasting interesting and gives me a bit of a loophole to promote this.
The frontpage/personal blog distinction has always been a bit confusing. I hope we can refactor it into something clearer at some point.
Maybe have a special “announcements” section on the frontpage?
Application deadlines have been extended to May 22! Feel free message me or Victor if you have any questions.
Hi Ryan, do you still plan for results to come out by May 27? And for those who are successful for the next stage to start June 6th etc.? (That’s what is says on the FAQ on the website still).
Yes, that is currently the plan. If we experience a massive influx of applications in the next two days it is possible this might slightly change, but I doubt it. We will work hard to keep to the announcement and commencement deadlines.
Hey, it’s now officially no longer May 27th anywhere, and I can’t find any announcements yet. How’s it going?Edit: Just got my acceptance letter! See you all this summer!