Analysis of AI Safety surveys for field-building insights
Introduction
To help grow the pipeline of AI safety researchers, I conducted a project to determine how demographic information (e.g. level of experience, exposure to AI arguments) affects AI researchers’ responses to AI safety. In addition, I examined additional AI safety surveys to uncover current issues preventing people from becoming AI safety researchers. Specifically, I analyzed the publicly-available data from the AI Impacts survey and also asked AI Safety Support and AGI Safety Fundamentals for their survey data (huge thank you to all three organizations). Below are my results, which I hope will be informative to future field-building efforts.
This work was done as part of the AI Safety Field-Building Hub; thanks to Vael Gates for comments and support. Comments and feedback are very welcome, and all mistakes are my own.
TLDR
AI Impacts
The following researchers tended to be more sympathetic to AI Alignment research:
Researchers with <=5 years of experience
Researchers who attended undergraduate universities in North America
Researchers who had thought extensively about HLMI (high-level machine intelligence) and/or about HLMI’s societal impacts
AI safety support
Surprisingly, academics were less concerned with funding compared to industry professionals.
AGISF
The vast majority of respondents said their post-course next steps would be seeking a job, internship, or fellowship. More “reading and understanding” was also cited heavily as was “applying, starting, or finishing” a PhD or other degree program.
Surprisingly, only 2⁄61 participants said they would seek funding.
Other sources
In speaking with researchers, there are wide views on everything from AGI timelines to the importance of AI alignment research. A few things were considered very helpful:
1) Explaining rather than convincing
2) Being open to questions and not dismissing criticism
3) Sharing existing research papers
Data
AI Impacts
The survey ran from June-August 2022 with participants representing a wide range of machine learning researchers. Note: error bars are standard error for proportions.
Summary
Academia was significantly more concerned about the existential risk as compared to industry. However, industry was significantly more concerned with human inability to control AI as compared to academia. Moreover, industry is somewhat more aggressive on HLMI timelines (not a huge surprise).
Junior researchers (<=5 years of experience) believe the risks of HLMI are greater than more seasoned professionals.
The more researchers had thought about HLMI, the more they believed in both an increase in prioritization and the existential risks.
The more researchers had thought about the societal implications of HLMI, the more they believed in both an increase in prioritization and the existential risks.
Researchers who attended undergraduate universities in North America had significantly higher probabilities of HLMI turning bad as compared to those who attended European or Asian universities.
By Industry
By Specific AI Field
Legend
CV = Computer Vision
DL = Deep learning
RL = Reinforcement learning
XAI = Trustworthy / Ethical AI / Fairness / AI Safety (short-term)
Application = Application / product (does not include autonomous driving) / data analysis
Note: This section is quite rough; Vael manually categorized these (not an expert in the field) very quickly, and Ash selected the first sub-speciality of the labels to make this graph.”
By Time in Career
By How Much Thought to HLMI
By How Much Thought to Social Impacts of Smarter-Than-Human Machine
By Undergraduate Continent
The full questions used in the survey for the graphs above were:≥ 5% chance that HLMI would be extremely bad
“Assume for the purpose of this question that HLMI will at some point exist. How probable do you find the overall extremely bad impact on humanity (e.g. human extinction), in the long run?”
Society should invest more/much more on AI safety research
“How much should society prioritize AI safety research, relative to how much it is currently prioritized?”
≥ 5% chance AI leading to bad outcomes
“What probability do you put on future AI advances causing human extinction or similarly permanent and severe disempowerment of the human species?”
≥ 5% chance humans can’t control AI leading to bad outcomes
“What probability do you put on human inability to control future advanced AI systems causing human extinction or similarly permanent and severe disempowerment of the human species?”
Error bars represent the standard error of the proportion. |
AI Safety Support
AI Safety Support conducted a broad survey to those interested in AI safety. One limitation to this data is that it’s from early 2021. Since then, new programs have come into existence that may slightly change the responses if they were provided today.
Summary
Surprisingly, academics were less concerned with funding compared to industry professionals.
Academics were much more interested in reading list, overview, and project/problem ideas compared to industry professionals.
Industry professionals were much more interested about jobs, internships, fellowships, and signaling value.
Note: The line between industry and academia can be blurry with overlap. However, the question specifically asked, “What best describes your current situation”.
What information are you missing?
Industry n=38, Academia n=27
What would help you?
Industry n=89, Academia n=31
Why not work on AI safety?
AGISF
The AGISF survey was announced on LessWrong and taken by a broad range of people.
Summary
The vast majority of respondents said their post-course next steps would be seeking a job, internship, or fellowship. More “reading and understanding” was also cited heavily as was “applying, starting, or finishing” a PhD or other degree program.
Surprisingly, only 2⁄61 said they would seek funding.
n=61
Additional
Vael Gates’s postdoc research
In their postdoctoral work, Vael Gates interviewed close to 100 researchers about their perceptions of AI safety. Vael recently shared their talk describing these results. Key findings include:
Very surprisingly, only 39% of researchers had heard of AI alignment.
In terms of whether they would work on AI alignment, there was a mixed response. Some said they think it is important and will work on it, or that they tried working on it in the past. Others felt they are unqualified to work on it, or that there needs to be more literature on the subject matter.
Participants had a very wide range of AGI timelines (similar to the AI Impacts findings).
Participants had a wide range of views regarding the alignment problem. From believing it is valid, to believing it will only be relevant a long time from now, to not placing any credence on it.
Marius Hobbhahn’s post
Marius Hobbhahn is a current PhD student who has spoken to more than 100 academics about AI Safety. He recently posted his takeaways.
His findings largely corroborate what we saw from the above sources (especially Vael Gates’ post-doc research) e.g.
Academics with less experience are more sympathetic to the AI safety and existential risk concerns.
In conversations with people, explain rather than try to convince.
Be open to questions and don’t dismiss criticism.
Existing papers that are validated within the academic community are super helpful starting points to send to researchers.
Limitations
These studies are not randomized control studies, nor have they been exposed to rigorous analytical examination. The knowledge repository is meant to provide directional understanding and subject to change as the community learns more.
Sample sizes on some questions are 100+, but on a few questions are smaller. I included everything with 30 or more responses.
As mentioned previously, the AI Safety Support survey was conducted in early 2021. Since then, new programs have come online and if the survey was to be repeated it might produce slightly different results.
Conclusion
I hope sharing my findings with the larger community will help build and grow the pipeline of AI safety and alignment research. This corpus of data is meant as a starting point. It will evolve as new programs and organizations sprout up to help solve some of these existing issues. AGI capability progress and funding will also be paramount in how views change going forward.
If you have any questions, please leave a comment below.
May I suggest adding an explanation for HLMI? I assume it means human level machine intelligence, but it might be useful to explicitly expand it.
Added the definition, thanks. It stands for “high-level machine intelligence”. AI Impacts goes on to describe it as “when unaided machines can accomplish every task better and more cheaply than human workers. Ignore aspects of tasks for which being a human is intrinsically advantageous, e.g. being accepted as a jury member. Think feasibility, not adoption.”