Hide table of contents

Introduction

To help grow the pipeline of AI safety researchers, I conducted a project to determine how demographic information (e.g. level of experience, exposure to AI arguments) affects AI researchers’ responses to AI safety. In addition, I examined additional AI safety surveys to uncover current issues preventing people from becoming AI safety researchers. Specifically, I analyzed the publicly-available data from the AI Impacts survey and also asked AI Safety Support and AGI Safety Fundamentals for their survey data (huge thank you to all three organizations). Below are my results, which I hope will be informative to future field-building efforts. 

 

This work was done as part of the AI Safety Field-Building Hub; thanks to Vael Gates for comments and support. Comments and feedback are very welcome, and all mistakes are my own. 

TLDR

  • AI Impacts
    • The following researchers tended to be more sympathetic to AI Alignment research:
      • Researchers with <=5 years of experience
      • Researchers who attended undergraduate universities in North America
      • Researchers who had thought extensively about HLMI (high-level machine intelligence) and/or about HLMI’s societal impacts
  • AI safety support
    • Surprisingly, academics were less concerned with funding compared to industry professionals.
  • AGISF
    • The vast majority of respondents said their post-course next steps would be seeking a job, internship, or fellowship. More "reading and understanding" was also cited heavily as was "applying, starting, or finishing" a PhD or other degree program. 
    • Surprisingly, only 2/61 participants said they would seek funding.
  • Other sources
    • In speaking with researchers, there are wide views on everything from AGI timelines to the importance of AI alignment research. A few things were considered very helpful: 
      • 1) Explaining rather than convincing
      • 2) Being open to questions and not dismissing criticism
      • 3) Sharing existing research papers 

Data

AI Impacts

The survey ran from June-August 2022 with participants representing a wide range of machine learning researchers. Note: error bars are standard error for proportions.

Summary

  • Academia was significantly more concerned about the existential risk as compared to industry. However, industry was significantly more concerned with human inability to control AI as compared to academia. Moreover, industry is somewhat more aggressive on HLMI timelines (not a huge surprise).
  • Junior researchers (<=5 years of experience) believe the risks of HLMI are greater than more seasoned professionals. 
  • The more researchers had thought about HLMI, the more they believed in both an increase in prioritization and the existential risks.
  • The more researchers had thought about the societal implications of HLMI, the more they believed in both an increase in prioritization and the existential risks.
  • Researchers who attended undergraduate universities in North America had significantly higher probabilities of HLMI turning bad as compared to those who attended European or Asian universities.
  • Public anonymized data

By Industry

By Specific AI Field

Legend

  • CV = Computer Vision
  • DL = Deep learning
  • RL = Reinforcement learning
  • XAI = Trustworthy / Ethical AI / Fairness / AI Safety (short-term)
  • Application = Application / product (does not include autonomous driving) / data analysis

 

 

Note: This section is quite rough; Vael manually categorized these (not an expert in the field) very quickly, and Ash selected the first sub-speciality of the labels to make this graph."

By Time in Career

By How Much Thought to HLMI

By How Much Thought to Social Impacts of Smarter-Than-Human Machine

By Undergraduate Continent

The full questions used in the survey for the graphs above were:

  • ≥ 5% chance that HLMI would be extremely bad
    • “Assume for the purpose of this question that HLMI will at some point exist. How probable do you find the overall extremely bad impact on humanity (e.g. human extinction), in the long run?”
  • Society should invest more/much more on AI safety research
    • “How much should society prioritize AI safety research, relative to how much it is currently prioritized?”
  • ≥ 5% chance AI leading to bad outcomes
    • “What probability do you put on future AI advances causing human extinction or similarly permanent and severe disempowerment of the human species?”
  • ≥ 5% chance humans can't control AI leading to bad outcomes
    • “What probability do you put on human inability to control future advanced AI systems causing human extinction or similarly permanent and severe disempowerment of the human species?”
Error bars represent the standard error of the proportion.

AI Safety Support

AI Safety Support conducted a broad survey to those interested in AI safety. One limitation to this data is that it's from early 2021. Since then, new programs have come into existence that may slightly change the responses if they were provided today.

Summary

  • Surprisingly, academics were less concerned with funding compared to industry professionals.
  • Academics were much more interested in reading list, overview, and project/problem ideas compared to industry professionals. 
  • Industry professionals were much more interested about jobs, internships, fellowships, and signaling value. 
  • Note: The line between industry and academia can be blurry with overlap. However, the question specifically asked, "What best describes your current situation".

What information are you missing?

Industry n=38, Academia n=27

What would help you?

Industry n=89, Academia n=31

Why not work on AI safety?

AGISF

The AGISF survey was announced on LessWrong and taken by a broad range of people.

Summary

  • The vast majority of respondents said their post-course next steps would be seeking a job, internship, or fellowship. More "reading and understanding" was also cited heavily as was "applying, starting, or finishing" a PhD or other degree program. 
  • Surprisingly, only 2/61 said they would seek funding. 

n=61

Additional

Vael Gates's postdoc research

  • In their postdoctoral work, Vael Gates interviewed close to 100 researchers about their perceptions of AI safety. Vael recently shared their talk describing these results. Key findings include:
    • Very surprisingly, only 39% of researchers had heard of AI alignment.
    • In terms of whether they would work on AI alignment, there was a mixed response. Some said they think it is important and will work on it, or that they tried working on it in the past. Others felt they are unqualified to work on it, or that there needs to be more literature on the subject matter.
    • Participants had a very wide range of AGI timelines (similar to the AI Impacts findings).
    • Participants had a wide range of views regarding the alignment problem. From believing it is valid, to believing it will only be relevant a long time from now, to not placing any credence on it. 

Marius Hobbhahn's post

  • Marius Hobbhahn is a current PhD student who has spoken to more than 100 academics about AI Safety. He recently posted his takeaways
  • His findings largely corroborate what we saw from the above sources (especially Vael Gates' post-doc research) e.g.
    • Academics with less experience are more sympathetic to the AI safety and existential risk concerns.
    • In conversations with people, explain rather than try to convince.
    • Be open to questions and don't dismiss criticism.
    • Existing papers that are validated within the academic community are super helpful starting points to send to researchers.

Limitations

  • These studies are not randomized control studies, nor have they been exposed to rigorous analytical examination. The knowledge repository is meant to provide directional understanding and subject to change as the community learns more.
  • Sample sizes on some questions are 100+, but on a few questions are smaller. I included everything with 30 or more responses.
  • As mentioned previously, the AI Safety Support survey was conducted in early 2021. Since then, new programs have come online and if the survey was to be repeated it might produce slightly different results.

 

Conclusion

I hope sharing my findings with the larger community will help build and grow the pipeline of AI safety and alignment research. This corpus of data is meant as a starting point. It will evolve as new programs and organizations sprout up to help solve some of these existing issues. AGI capability progress and funding will also be paramount in how views change going forward.

If you have any questions, please leave a comment below.



 

Comments7


Sorted by Click to highlight new comments since:

There's a minor mistake in the post

AI Safety Support conducted surveys to people they had worked with previously.

Actually, most people where not people we engaged with before. We spread the survey as wide as we could. 

Half of them came from Rob Miles audience, or that's my estimate based on timing of responses.


But thanks for analysing our data! 

I never got around to it, since I had to leave AISS a few months later, under pretty bad circumstances (which was not the fault of anyone working for AISS!)

 

Another interesting result of that survey was that a substantial fraction of applicants wanted some type of help or resource that did already existed but which they were unaware of. I can't remember now what percentage of applicants this was, but I think it was something in the range 20%-70%

I wrote about it here:
AI Safety Career Bottlenecks Survey Responses Responses 

Hi Linda - Thanks for the clarification on the source of your data. Updated with the corrected information.

Thanks for writing this up. Very interesting and helpful!

Wonderful to see an analysis of all this even more wonderful data coming out of the programs! We're also happy to share our survey results (privately and anonymized) with people interested:

  • Demographic, feedback, AIS attitudes before and after an Alignment Jam (for the first and second versions)
  • A 70 respondent in-depth survey on attitudes towards AI risk arguments (basically, "what converted you")
  • An AI safety pain points survey that was also done in-person with several top AIS researchers
  • An AI safety progress survey about P(doom)s etc.

We also have a few other surveys but the ones above are probably most interesting in the context of this post.

Hi, I am interested in seeing this data! Thanks!

Very interested!

Very confused why people have downvoted here? Please explain why, if possible. And context: I've also started a private chat about the data with Peter and Ash.

Curated and popular this week
 ·  · 5m read
 · 
This work has come out of my Undergraduate dissertation. I haven't shared or discussed these results much before putting this up.  Message me if you'd like the code :) Edit: 16th April. After helpful comments, especially from Geoffrey, I now believe this method only identifies shifts in the happiness scale (not stretches). Have edited to make this clearer. TLDR * Life satisfaction (LS) appears flat over time, despite massive economic growth — the “Easterlin Paradox.” * Some argue that happiness is rising, but we’re reporting it more conservatively — a phenomenon called rescaling. * I test rescaling using long-run German panel data, looking at whether the association between reported happiness and three “get-me-out-of-here” actions (divorce, job resignation, and hospitalisation) changes over time. * If people are getting happier (and rescaling is occuring) the probability of these actions should become less linked to reported LS — but they don’t. * I find little evidence of rescaling. We should probably take self-reported happiness scores at face value. 1. Background: The Happiness Paradox Humans today live longer, richer, and healthier lives in history — yet we seem no seem for it. Self-reported life satisfaction (LS), usually measured on a 0–10 scale, has remained remarkably flatover the last few decades, even in countries like Germany, the UK, China, and India that have experienced huge GDP growth. As Michael Plant has written, the empirical evidence for this is fairly strong. This is the Easterlin Paradox. It is a paradox, because at a point in time, income is strongly linked to happiness, as I've written on the forum before. This should feel uncomfortable for anyone who believes that economic progress should make lives better — including (me) and others in the EA/Progress Studies worlds. Assuming agree on the empirical facts (i.e., self-reported happiness isn't increasing), there are a few potential explanations: * Hedonic adaptation: as life gets
 ·  · 38m read
 · 
In recent months, the CEOs of leading AI companies have grown increasingly confident about rapid progress: * OpenAI's Sam Altman: Shifted from saying in November "the rate of progress continues" to declaring in January "we are now confident we know how to build AGI" * Anthropic's Dario Amodei: Stated in January "I'm more confident than I've ever been that we're close to powerful capabilities... in the next 2-3 years" * Google DeepMind's Demis Hassabis: Changed from "as soon as 10 years" in autumn to "probably three to five years away" by January. What explains the shift? Is it just hype? Or could we really have Artificial General Intelligence (AGI)[1] by 2028? In this article, I look at what's driven recent progress, estimate how far those drivers can continue, and explain why they're likely to continue for at least four more years. In particular, while in 2024 progress in LLM chatbots seemed to slow, a new approach started to work: teaching the models to reason using reinforcement learning. In just a year, this let them surpass human PhDs at answering difficult scientific reasoning questions, and achieve expert-level performance on one-hour coding tasks. We don't know how capable AGI will become, but extrapolating the recent rate of progress suggests that, by 2028, we could reach AI models with beyond-human reasoning abilities, expert-level knowledge in every domain, and that can autonomously complete multi-week projects, and progress would likely continue from there.  On this set of software engineering & computer use tasks, in 2020 AI was only able to do tasks that would typically take a human expert a couple of seconds. By 2024, that had risen to almost an hour. If the trend continues, by 2028 it'll reach several weeks.  No longer mere chatbots, these 'agent' models might soon satisfy many people's definitions of AGI — roughly, AI systems that match human performance at most knowledge work (see definition in footnote). This means that, while the compa
 ·  · 4m read
 · 
SUMMARY:  ALLFED is launching an emergency appeal on the EA Forum due to a serious funding shortfall. Without new support, ALLFED will be forced to cut half our budget in the coming months, drastically reducing our capacity to help build global food system resilience for catastrophic scenarios like nuclear winter, a severe pandemic, or infrastructure breakdown. ALLFED is seeking $800,000 over the course of 2025 to sustain its team, continue policy-relevant research, and move forward with pilot projects that could save lives in a catastrophe. As funding priorities shift toward AI safety, we believe resilient food solutions remain a highly cost-effective way to protect the future. If you’re able to support or share this appeal, please visit allfed.info/donate. Donate to ALLFED FULL ARTICLE: I (David Denkenberger) am writing alongside two of my team-mates, as ALLFED’s co-founder, to ask for your support. This is the first time in Alliance to Feed the Earth in Disaster’s (ALLFED’s) 8 year existence that we have reached out on the EA Forum with a direct funding appeal outside of Marginal Funding Week/our annual updates. I am doing so because ALLFED’s funding situation is serious, and because so much of ALLFED’s progress to date has been made possible through the support, feedback, and collaboration of the EA community.  Read our funding appeal At ALLFED, we are deeply grateful to all our supporters, including the Survival and Flourishing Fund, which has provided the majority of our funding for years. At the end of 2024, we learned we would be receiving far less support than expected due to a shift in SFF’s strategic priorities toward AI safety. Without additional funding, ALLFED will need to shrink. I believe the marginal cost effectiveness for improving the future and saving lives of resilience is competitive with AI Safety, even if timelines are short, because of potential AI-induced catastrophes. That is why we are asking people to donate to this emergency appeal