This is a Draft Amnesty Week draft. It may not be polished, up to my usual standards, fully thought through, or fully fact-checked. |
This draft lacks the polish of a full post, but the content is almost there. The kind of constructive feedback you would normally put on a Forum post is very welcome. |
I wrote most of this last year. I also think I’m making a pretty basic point and don’t think I’m articulating it amazingly, but I’m trying to write more and can imagine people (especially newer to EA) finding this useful - so here we go
Last week[1] I was at an event with a lot of people relatively new to EA - lots of them had recently finished the introductory fellowship. Talking through their plans for the future, I noticed that many of them used the concept ‘personal fit’ to justify their plans to work on a problem they had already found important before learning about EA.
They would say they wanted to work on combating climate change or increasing gender equality, because
- They had studied this and felt really motivated to work on it
- Therefore, their ‘personal fit’ was really good for working on this topic
- Therefore surely, it was the highest impact thing they could be doing.
I think a lot of them were likely mistaken, in one or more of the following ways:
- They overestimated their personal fit for roles in these (broad!) fields
- They underestimated the differences in impact between career options and cause areas
- They thought that they were motivated to do the most good they could, but in fact they were motivated by a specific cause
To be clear: the ideal standard here is probably unattainable, and I surely don’t live up to it. However, if I could stress one thing, it would be that people scoping out their career options could benefit from first identifying high-impact career options, and only second thinking about which ones they might have a great personal fit for - not the other way around.
- ^
This was last year
I agree narrowly with the idea that the option people could have the most impact in isn't necessarily the thing they like.
But the option a person could have the most early career impact in isn't necessarily what's on EA-recommended lists either:
And they're not entirely wrong that doing stuff they're already engaged with is a "fit" even if they don't have any specialist skills in that field because
So whilst it's entirely possible that people are engaging in motivated reasoning, there's also reason to be cautious about going the other way, and deferring too much to impactful cause area recommendations, or assuming that people can't be more impactful in cause areas the likes of OpenPhil think are not neglected.