Hide table of contents

Help support the next generation of AI safety talent

AI Safety Global Society is a nonprofit organisation helping people transition into high-impact research careers. We run upskilling programs that train and mentor early to mid-career researchers working to reduce existential risks from transformative AI. 

We’re looking for experienced people to support this work as teaching assistants, research leads, and facilitators. We are currently applying for funding. This is an expression of interest for anyone who would like to be involved in the first iteration of our programs.

Open Roles

Research Leads

Program: AI Control Research Fellowship


Duration: 8 weeks
Approx dates: TBC
Time commitment: ~10 hours/week
Format: Remote, part-time

You will support early to mid-career researchers in researching AI control topics, such as monitoring, control evaluations, or capability control.

Responsibilities:

  • Lead a small research group (3–5 fellows)
  • Help design and scope projects
  • Coordinate with the AISGS team and fellow mentors
  • Review and guide research outputs
     

Ideal background:

Facilitators

Program: AI Control Reading Group


Based on: Louis Thomson Syllabus, with reference to Ram Potham’s recent literature review and Ryan Greenblatt’s overview of areas in control work.


Duration: 6 weeks (1 session/week)
Approx dates: TBC
Time commitment: 2–5 hours/week
Format: Remote  (Weekly 2-hour discussion sessions + ~2 hours prep). You can find the curriculum structure here

Responsibilities:

  • Lead weekly 2-hour discussions
  • Explain technical concepts clearly and help connect ideas across the syllabus
  • Encourage thoughtful participation and collaborative learning

Ideal background:

  • Familiarity with at least 2–4 topics from Ram Potham’s AI control methods literature review
  • Background in research, engineering, or technical AI safety (does not need to be in AI control specifically)
  • Comfortable explaining technical concepts and facilitating discussion

     

Teaching Assistants

Program: Alignment Research Fellowship 


Duration: 8 weeks
Approx dates: TBC
Based on: ARENA curriculum

The Alignment Research Fellowship helps participants build technical foundations for alignment research, covering ML basics, transformers, and mechanistic interpretability. Final projects explore reinforcement learning, evaluation methods or mechanistic interpretability with NDIF.

Responsibilities:

  • Guide participants through technical exercises
  • Provide implementation feedback
  • Help scope and supervise final projects

Ideal background:

  • Strong PyTorch skills
  • Familiarity with transformers and mechanistic interpretability
  • Experience with RL or evaluations is a plus
  • Prior participation in ARENA or similar programs is a plus

If you're unsure whether you're a good fit, err on the side of applying.

Why work with us?

Advance the field of AI safety: Contribute to a growing ecosystem of researchers working to reduce risks from transformative AI.
Give back: Help talented researchers access high-quality courses, enabling them to pursue impactful careers in AI safety.
Join a growing community: Connect with researchers focused on mitigating risks from transformative AI.

Other benefits include:

  • Co-authorship on research papers (where applicable)
  • Access to structured curricula and compute resources
  • Letters of recommendation

Interested?
We welcome expressions of interest here or feel free to contact us at contact@aisglobalsociety.org

14

0
0

Reactions

0
0

More posts like this

Comments
No comments on this post yet.
Be the first to respond.
Curated and popular this week
Relevant opportunities