Griffin Young

Student @ Stanford University
Pursuing a graduate degree (e.g. Master's)
Seeking work
10Joined Jul 2022

Bio

I'm a coterm in Computer Science at Stanford, which is a graduate degree. I did my undergrad here as well in Symbolic Systems.

I recently started doing the 80,000 hours career planning course. 

I'm most compelled by suffering-based ethics, though I still find negative utilitarianism unsatisfactory on a number of edge cases. This makes me less worried about X-risks and by extension less long-termist than seems to be the norm.

My shortlist of cause areas is the following:

  1. Global priorities research (depends on 2 and 4 below)
  2. Factory farming (depends on 3 below)
  3. Mental health
  4. Painful medical conditions
  5. Great power conflict
  6. Biorisk
  7. Climate change
  8. AI risk (pretty confident it will happen, and decently confident it will happen somewhat soon; depends on 1 below)

though I remain uncertain, especially about the following things:

  1. Potential badness of AI
  2. Scale of untapped cause areas/when we will reach a “saturation point” of finding the best areas
  3. Relative amount of animal suffering
  4. Trajectory of future EA funding

How others can help me

I am looking for employment once I graduate, which will be in December 2022.

How I can help others

I have experience teaching AI, so I can help answer questions about some of the fundamentals.

Comments
3

What do you think would be most people's cutoff? My guess would be that most people see the two types of suffering as qualitatively different such that no amount of insect suffering is comparable to human suffering.

How many AI Safety researchers would be enough? 80k emphasizes the fact that there are only 300 people working on this full-time, meaning that the problem is extremely neglected. How many people would have to be working on this problem for it to no longer be considered neglected?