(I am writing this post under a pseudonym because I don’t want potential future non-EA employers to find this with a quick google search. Initially my name could be found on the CV linked in the text, but after this post was shared much more widely than I had expected, I got cold feet and removed it.)
In the past 12 months, I applied for 20 positions in the EA community. I didn’t get any offer. At the end of this post, I list all those positions, and how much time I spent in the application process. Before that, I write about why I think more posts like this could be useful.
Please note: The positions were all related to long-termism, EA movement building, or meta-activities (e.g. grant-making). To stress this again, I did not apply for any positions in e.g. global health or animal welfare, so what I’m going to say might not apply to these fields.
Costs of applications
Applying has considerable time-costs. Below, I estimate that I spent 7-8 weeks of full-time work in application processes alone. I guess it would be roughly twice as much if I factored in things like searching for positions, deciding which positions to apply for, or researching visa issues. (Edit: Some organisations reimburse for time spent in work tests/trials. I got paid in 4 of the 20 application processes. I might have gotten paid in more processes if I had advanced further).
At least for me, handling multiple rejections was mentally challenging. Additionally, the process may foster resentment towards the EA community. I am aware the following statement is super in-accurate and no one is literally saying that, but sometimes this is the message I felt I was getting from the EA community:
“Hey you! You know, all these ideas that you had about making the world a better place, like working for Doctors without Borders? They probably aren’t that great. The long-term future is what matters. And that is not funding constrained, so earning to give is kind of off the table as well. But the good news is, we really, really need people working on these things. We are so talent constraint… (20 applications later) … Yeah, when we said that we need people, we meant capable people. Not you. You suck.”
Why I think more posts like this would have been useful for me
Overall, I think it would have helped me to know just how competitive jobs in the EA community (long-termism, movement building, meta-stuff) are. I think I would have been more careful in selecting the positions I applied for and I would probably have started exploring other ways to have an impactful career earlier. Or maybe I would have applied to the same positions, but with less expectations and less of a feeling of being a total loser that will never contribute anything towards making the world a better place after being rejected once again 😊
Of course, I am just one example, and others will have different experiences. For example, I could imagine that it is easier to get hired by an EA organisation if you have work experience outside of research and hospitals (although many of the positions I applied for were in research or research-related).
However, I don’t think I am a very special case. I know several people who fulfil all of the following criteria:
- They studied/are studying at postgraduate level at a highly competitive university (like Oxford) or in a highly competitive subject (like medical school)
- They are within the top 5% of their course
- They have impressive extracurricular activities (like leading a local EA chapter, having organised successful big events, peer-reviewed publications while studying, …)
- They are very motivated and EA aligned
- They applied for at least 5 positions in the EA community and got rejected in 100% of the cases.
I think I also fulfil all these criteria. Here is my CV roughly at the time when I was doing the applications. It sports such features as ranking 16th out of around 6000 German medical students, and 8 peer-reviewed publications while studying.
Without further ado, here are all the ...
Positions I got rejected from in the last 12 months
I also include the stage that I was rejected at and how much time I had invested in the application process (Mostly work tests, but also researching organisations, adapting personal statements, preparing for interviews. I am counting “lost productivity” here, so I am also counting travel time weighted at around 50%).
Position – how far I got– how much time I invested in the application
Chief of Staff at Will MacAskill’s Office – stage 2/2: didn’t get an offer after 2 days worktrial – 32 h
OpenPhil Research Analyst – stage 2/4 (?): rejected after conversation notes work test – 22 h
OpenPhil biosecurity early career researcher grant – stage 1/1: no grant – 40 h
EA grants evaluator (CEA) – stage 2/X: rejected after first interview – 7 h
FHI Research scholar programme – stage 2/3: rejected after second work test – 50 h
Effective giving uk researcher – Stage 3/3 (?): no offer after what I think was the final interview – 15 h
LEAN manager – Stage 2/?: rejected after work test – 6 h
CEA operations specialist – stage 3/?: rejected after interview – 9 h
CEA local group specialist – stage 2/?: rejected after work-test – 12 h
2x FHI academic project manager (GovAI and research scholar programme) – Stage 2/2: no offer after final interview – 10 h each
Toby Ord research assistant – Stage 2/?: rejected after work test - 12 h
Center for Health Security research analyst – initiative application and interview, but they decided not to hire at all – 10 h
Nuclear Threat initiative researcher – initiative application, never heard back – 1 h
CSER biosecurity postdoc – stage 1/? – 3 h
CSER academic project manager – stage 1/? – 2 h
GPI Head of Research Operations – Stage 4/4: No offer after in-person work trial – 32 h
And here are additional positions I applied for but then did not complete the application process:
COO Ought – stopped following up after first stage because of visa issues – 4 h (but much more if you count me researching said visa issues)
Researcher Veddis – I decided not to go on the final stage work trial – 6 h
Program Manager/Investigator at BERI– I decided not to do the final stage work test – 15 h
Adding some more data from my own experience last year.
Personally, I'm glad about some aspects of it and struggled with others, and there are some things I wish I had done differently, at least in hindsight. But here I just mean to quickly provide data I have collected anyway in a 'neutral' way, without implying anything about any particular application.
Total time I spent on 'career change' in 2018: at least 220h, of which at least about 101h were for specific applications. (The rest were things like: researching job and PhD opportunities; interviewing people about their jobs and PhD programs; asking people I've worked with for input and feedback; reflection before I decided in January to quit my previous job at the EA Foundation by April.) This does neither include 1 week I spent in in San Francisco to attend EAG SF and during which I was able to do little other work nor 250h of self-study that seems robustly useful but which I might not have done otherwise. (Nor 6 full weeks plus about 20h afterwards I spent doing an internship at an EA org, which overall I'm glad I did but might not have done otherwise.)
- Open Phil Research Analyst - rejected af
... (read more)One thing that might be worth noting: I was only able to invest that many resources because of things like (i) having had an initial runway of more than $10,000 (a significant fraction of which I basically 'inherited' / was given to me for things like academic excellence that weren't very effortful for me), (ii) having a good relationship to my sufficiently well-off parents that moving back in with them always was a safe backup option, (iii) having access to various other forms of social support (that came with real costs for several underemployed or otherwise struggling people in my network).
I do think current conditions mean that we 'lose' more people in less comfortable positions than we otherwise would.
+1 to noting that the current recruitment configuration strongly favors elite (& highly privileged) applicants.
Yeah, this is one reason Open Phil pays people for doing our remote work tests, so that people who don't happen to have runway/similar can still go through our process. Possibly more EA orgs should do this if they aren't already.
I'd like to make this into a norm, but it does also pose a barrier for funding constrained EA organizations by increasing the costs of hiring.
I think it's fine to be a "norm, if you can afford it."
If you can't afford it, doesn't that suggest that earning to give might not be such a bad choice after all?
Yes. Earning to give is a good choice and I've not suggested otherwise.
(Peter has been one of several people continuing to argue "earning to give is undervalued, most orgs could still do useful things with more funding".)
Just a thank you for sharing, it can be scary to share your personal background like this but it's extremely helpful for people looking into EA careers.
I think there are at least two effects where the world loses impact: (i) People in less privileged positions not applying for EA jobs; sometimes one of these would actually have been the best candidate. (ii) More speculatively (in the sense that I can't point to a specific example, though my prior is this effect is very likely to be non-zero), people in less privileged positions might realize that it's not possible for them to apply for many of the roles they perceived to be described as highest-impact and this might reduce their EA motivation/dedication in general, and make them feel unwelcome in the community.
I emphatically agree that them taking another potentially impactful job is positive. In fact, as I said in another comment, I wish there was more attention on and support for identifying and promoting such jobs.
Something that seems to be missing from this (very valuable) conversation is that many people also spend months looking for non-EA jobs that they have a personal fit for. I'm mainly aware of people with science PhDs, either applying for industry jobs or applying for professorships. It is not uncommon for this to be a months long process with multiple 10s of applications, as being reported here for EA job searching. The case of where this goes faster in industry jobs tends to be because the applicant is well established as having a key set of skills that a company needs and/or a personal network connection with people involved in hiring at the company. Some academics get lucky just applying for a few professorships, but others apply to 50+ jobs, which easily takes 100+ hours, perhaps many more. And in both cases you spend lots of time over the preceding years learning about the job search process, how to write cover letters, teaching statements, etc.
I really like the specific numbers people are posting. I'll add my own (rough estimates) from the ~5 months I spent applying to roles in 2018.
Context: In spring 2018, I attended an event CEA ran for people with an interest in operations, because Open Phil referred me to them; this is how I wound up deciding to apply to most of the roles below. Before attending the operations event, I'd started two EA groups, one of which still existed, and spent ~1 year working 5-10 hours/week as a private consultant for a small family foundation, doing a combination of research and operations work. All of the below experiences were specific to me; others may have gone through different processes based on timing, available positions, prior experience with organizations, etc.
- CEA (applied to many positions, interviewed for all of them at once, didn't spend much additional time vs. what I'd have done if I just applied to one)
- ~4 hours of interview time before the work trial, including several semi-casual conversations with CEA staff at different events about roles they had open.
- ~2-hour work trial task, not very intense compared to Open Phil's tasks
- 1.5-week work trial at CEA; th
... (read more)I really wish we (as an EA community) didn't work so hard to accidentally make earning to give so uncool. It's a job that is well within the reach of anyone, especially if you don't have unrealistic expectations of how much money you need to make and donate to feel good about your contributions. It's also a very flexible career path and can build you good career capital along the way.
Sure talent gaps are pressing, but many EA orgs also need more money. We also need more people looking to donate, as the current pool of EA funding feels very over-concentrated in the hands of too few decision-makers right now.
I also wish we didn't accidentally make donating to AMF or GiveDirectly so uncool. Those orgs could continually absorb the money of everyone in EA and do great, life-saving work.
(Also, not to mention all the career paths that aren't earning to give or "work in an EA org"...)
+1 for pointing out the hazard of having funding concentrated in the hands of a very few decision makers
>Also, not to mention all the career paths that aren't earning to give or "work in an EA org"
While I share your concern about the way earning to give is portrayed, I think this issue might be even more pressing.
FWIW, without having thought systematically about this, my intuition is to agree. I'd be particularly keen to see:
My gut feeling is that in many places we could do a better job at utilizing skills and experiences people can get pretty reliably in the for-profit world, academia, or from other established 'institutions'.
I'm aware this is happening to some extent already, e.g. GPI trying to interface with academia or 80K's guide on US policy. I think both are great!
NB this is different from the idea that there are many other career paths that would be high-impact to stay in indefinitely. I think this is also true, but at least if one has a narrow focus on the long-term future I feel less sure if there are 'easy wins' left here.
(An underlying disagreement here might be: Is this feasible, or are we just too much bottlenecked by something like what Carrick Flynn has called 'disentanglement'. Very ... (read more)
Nice point.
'I also wish we didn't accidentally make donating to AMF or GiveDirectly so uncool.'
This reminds me of the pattern where we want to do something original, so we don't take the obvious solution.
I'm really glad that's been your experience and I acknowledge that maybe my experience isn't typical.
My experience has been more pessimistic. Honestly, I usually encounter conversations that feel more like this:
Bob: “Hi, I can donate $10,000 a year to the EA movement. GiveWell says that could save 4-5 lives a year, and it’s quite possible we could even find better giving opportunities than GiveWell top charities. This is super exciting!”
Alice: “Pff, $10K/yr isn’t really that much. We don’t need that. You should do direct work instead.”
Bob: “Ok, how about I research biosecurity?”
Alice: “Nah, you’d probably mess that up. We should just let FHI handle that. We can’t talk about this further because of infohazards.”
...Obviously this is dramatized for effect, but I've never seen a community so excited to turn away money.
In addition to what Peter describes, if we do a simple content analysis of forum threads or blog posts in the last 3 or so years, ETG feels like it's become invisible. Long term EAs like you and me most likely do still think it's cool because when we became EAs it was a huge part of it and probably a big part of what drew us in (in my case, certainly - I became an EA the year GWWC was launched). But that doesn't mean that this is the subtext that newer EAs are getting. I feel like the opposite is true, and I find that deeply concerning.
Oof, 8 weeks of effort to get 0/20 positions is pretty brutal. It's easy to see how that would feel like your "Hey you!…" paragraph. And while I suspect you're a bit of an outlier in time spent and positions applied for, I also think you're pointing at something true about the current situation re: job openings at EA-motivated employers, as evidenced by how many upvotes this post has gotten, some of the comments on this page, and the data I've got as a result of managing Open Phil's 2018 recruitment round of Research Analysts, during which we had to say "no" to tons of applicants with quite impressive resumes.
I've been writing up some reflections on that recruiting round, which I hope to share soon. One of my takeaways is something like "The base of talent out there is strong, and Open Phil's current ability to deploy it is weak." In that way we might be an extreme opposite of Teach for America, and I suspect many other EA-motivated orgs are as well.
Anyway, I plan to say more on these topics when I share my "reflections" post, but in the meantime I just want to say I'm sorry that you spent so much time applying to EA orgs and got no offers. Also, setting the time investment aside,
... (read more)BTW my "reflections on the 2018 RA recruiting round" post is now up, here.
MIRI (and other EA orgs, I'd wager) would strongly second "we don't think it's a waste of our time to process applications. If we don't have the staff capacity to process all the applications we receive, we can always just drop a larger fraction of applicants at each stage."
I second the rest of Luke's comment too. That run of applications sounds incredibly rough. The account above makes me wonder if we could be doing a better job of communicating expectations to people applying for jobs early in the process. It's much, much easier to avoid setting misleadingly low or misleadingly high expectations when the information can be personalized and there's an active back-and-forth, vs. in a blog post.
Hey :-)
Regarding me being a bit of an outlier: Yes, I think so as well. I personally don't know anyone who applied for quite as many positions. I still don’t think I am a *very* special case. I also got several private messages in response to this post, of people saying they had made similar experiences.
Regarding compensation: I was lucky enough to have decent runway, so the financial aspect wasn't crucial for me, so I just forgot including it. I will edit that now.
Of course, the one who writes the post about it is likely to be the outlier rather than the median.
Just want to add my voice to the many who have already said: thank you for sharing this. It must have taken some courage.
My own experience has been similar (though I'm far less qualified than the OP).
I've recently graduated from one of the top ~10 universities worldwide, after investing heavily in EA throughout my studies. While a student, EA was the biggest thing in my life. I read a lot, and several of my EA peers told me I stood out as particularly well-informed about EA topics, especially long-termist ones. Eventually I contributed some of my own research too. I also invested enormous amounts of time in student EA projects. Many people, including ones I thought well-informed about the talent landscape, fully expected that I would go work for an 'EA organisation'. Naively, I believed it too.
Over the last seven months, I've made over 20 unsuccessful job applications (I keep a spreadsheet). This has increased the severity of my depression and anxiety. Over time, I began to shed my identity as an EA, no doubt as a self-defence mechanism. Now I'm very disillusioned about my ability to contribute to the long-termist project.
Thanks for sharing. As someone who spends a lot of time trying to fill EA meta/longtermist talent gaps — e.g. by managing Open Phil RA recruiting, helping to match the strongest applicants we don't hire to other openings, and by working on field-building in AI strategy/policy (e.g. CSET) — hearing stories like yours is unnerving.
What changes to the landscape, or hiring processes, or whatever, do you think would've made the most difference in your case?
I'm also curious to hear your reaction to my comment elsewhere about available paths:
... (read more)(My answers might be very different to the ones that anonymousthrowaway might give).
Even if not a direct answer to your question, maybe it helps to illuminate the dynamics that contributed towards me making that many applications before moving on to other EA things. I personally did NOT think that jobs at EA organisations had clearly higher expected value than my other top options (which were working in biosecurity basically anywhere, or upskilling in machine learning via a master's degree). (There were very few exceptions, like Chief of Staff for Will or RA at OpenPhil, which I thought were outstandingly good).
Then why did I apply to so many positions?
1. I thought EA organisations were really talent-starved and needed me. I also thought that it would be easy to get a job. This is clearly my fault, and I think there would have been ways for me to reach a better knowledge of the situation. But I went to several EAGs, talked to dozens of people, read most of 80000 hours advice, and I think it was definitely quite easy to come away from these with the impression I had.
2. I got pretty far in a few application processes, so that encouraged me to continue applying.
3. Somehow, EA ... (read more)
Regarding biosecurity roles from 80,000 Hours: While they don't seem to have any jobs in that field being actively promoted on their job board, they do maintain a list of "organizations we recommend" in biosecurity on this page, which might be useful for getting a head start on learning about these orgs' work.
Thanks, that all makes sense to me. Will think more about this. Also still curious to hear replies from others here.
[deleted]
Thanks for +1ing the above comment. I'd be keen to hear your reply to this comment, too.
Just wanted to say I appreciate this post and I feel you. I had a very similar feeling when applying to receive coaching from 80K. I think they've improved their message since I applied, but at the time I didn't fully realize the type of person they were taking was literally the going to Harvard, in the middle of a machine learning Phd type.
I also feel like generally there's a huge under-utilization of talent in EA. I see so many cases of smart, motivated people who may not have literally the best resume ever, but could be extremely helpful *somewhere*. It's strange considering the emphasis on talent constraints.
I don't have an easy solution unfortunately and it does seem to be a really hard problem, but the Teach for America like funnel that I believe Nick Beckstead recommended and was brought up in a recent EA forum post seems really promising to me.
In a similar vein, I don't mean to pick on 80K - of course they do great work - but I'm confused why they don't just hire more career coaches. I realize they want to remain lean and have a really strong team, but I think they only have 2 coaches. That number is going to leave a lot of promising EAs wandering in the dark for a while.
p.s. I also know of an EA that applied to EA orgs for a long time and got rejected by all of them. They now work at one of the FAANG companies.
Hi Kevin,
Howie from 80k here. Note that I'm just one staff member and don't speak on behalf of the whole team, but I wanted to give some thoughts on your comment.
First of all, I appreciate the feedback and I’m sorry to hear that you had a frustrating experience applying for coaching from 80k. Unfortunately, we can only work with a small fraction of the people who apply. Our current coaching page tries to make this clear and I apologize if we’ve done a bad job communicating this (either currently or in the past).
I understand that this situation might be particularly frustrating in light of the EA community’s recent emphasis on ‘talent gaps.’ We think this term has ended up causing a lot of confusion and it’s partly our fault, so we wrote a post back in November that tries to clarify our thinking on this issue.
In particular, we believe that very specific skill sets are the biggest constraint for most of our priority problems, as opposed to ‘talent’ in general. To address this, we’ve increased our focus on what we call ‘priority paths,’ which are the best ways we know of for people to use their careers to free up these specific constraints. Our current priority in coaching ... (read more)
Hi Howie,
I really wasn't expecting a reply from an 80K staff member so I really appreciate taking time out to give myself and other readers more context. FWIW, I think generally the moves you describe make sense.
On "In our experience, these people aren't 'wandering in the dark;'" - this is just anecdotal evidence but I'd like to push back on that a bit: That phrase might be a bit of an exaggeration but I think personally in my own case and in a number of EAs I've come across, a 20-60 min session with an expert could have helped substantially, perhaps shaving off years of semi-time wasting.
For example in my own case, I had read many of 80K's articles but hadn't come across the more advanced advice of downgrading broad-based advocacy for the very top causes. I spent ~1 year working on media advocacy without fully knowing that - I'm not trying to blame 80K here, just that a short conversation with an expert who could really push back on my plan and point me in a particular direction I think would have been really useful. The specific advice I'm talking about is here: https://80000hours.org/articles/extinction-risk/#ways-t... (read more)
I think part of what might be driving the difference of opinion here is that the type of EAs that need a 45 minute chat are not the type of EAs that 80k meets. If you work at 80k, you and most of the EAs you know: probably have dozens of EA friends, have casual conversations about EA, pick up informal knowledge easily, and can talk out your EA ideas with people who can engage. But the majority of people who call themselves EA probably don't have many if any friends who work at EA organizations, donate lots, provide informal knowledge of EA, or who can seriously help you figure out how to have a high impact career.
A 45 minute discussion can therefore do a lot more good for someone outside the EA social circle than for someone who has friends who can have this conversation with them.
I agree that a quick and decisive input from someone very knowledgeable about EA and the topic involved would be very useful and save a lot of time and indecision for people evaluating career options.
I think we can provide a bit of this though through more engaged online communities around given topic areas. Not nearly as good as in person talks but people can at least get some general feedback on career ideas. I'm hoping to host an event later this year that will gather people interested in a cause area and use that as a catalyst to form a more cohesive online community. As far as I can tell (and in my experience) people tend not to engage much in an online community if they don't really know the people well. Though it's definitely true that some people are more than happy to engage with people they don't know.
I don't know how this could move forward but it seems like someone could potentially make a difference by engineering Facebook or Slack groups focused on certain cause areas to be more active places for general discussion and career advice. This would be so helpful for people who lack close contact with knowledgeable people in EA or within their cause area.
Strongly agreed. I really like Raemon's analysis why it's so hard to get EA careers: we're network constrained. [This isn't exactly how he frames it, more my take on his idea.]
Right now, EA operates very informally, relying heavily on the fact that the several hundred people working at explicitly EA orgs are all socially networked together to some degree. This social group was significantly inherited from LessWrong and Bay Area rationalism, and EA has had great success in co-opting it for EA goals.
But as EA grows beyond its roots, more people want in, and you can't have a social network of ten thousand, let alone a million. So we have two options: (a) increase the bandwidth of the social network, or (b) stop relying so much on the social network.
(a) increasing bandwidth looks like exactly what you're talking about: create ways for newcomers to EA to make EA friends, develop professional relationships with EAs, etc., by creating better online platforms and in person groups.
(b) not relying on personal relationships looks like becoming more corporate, relying on traditional credentials, scaling up until people actually stand a strong chance of landing jobs via open application, etc.
(a) seems to have clear benefits with no obvious harms, as long as it can be done, so it seems very much worth it for us to try.
Some related half-baked thoughts:
[Epistemic status: I appreciate that there are people who've thought about the EA talent landscape systematically and have access to more comprehensive information, e.g. perhaps some people at 80K or people doing recruiting for major EA orgs. I would therefore place significantly more weight on their impressions. I'm not one of these people. My thoughts are based on (i) having talked 10-100 hours with other EAs about related things over the last year, mostly in a non-focussed way, (ii) having worked full-time for 2 EA organizations (3 if one counts a 6-week internship), (iii) having hired 1-5 people for various projects at the Effective Altruism Foundation, (iv) having spent about 220h on changing my career last year, see another comment. I first heard of EA around October 2015, and have been involved in the community since April 2016. Most of that time I spent in Berlin, then over last summer and since October in Oxford.]
- I echo the impression that several people I've talked to - including myself - were or are overly focussed on finding a job at a major EA org. This applies both in terms of time spent and number of applications submit
... (read more)Not sure I follow the part about how the kind of thing described in the original post makes you "more reluctant to introduce new people into the EA community." There are lots of exciting things for EAs to do besides "apply to one or more of the 20 most competitive jobs at explicitly EA-motivated employers," including "keep doing what you're doing and engage with EA as an exciting hobby" and "apply to key positions in top-priority cause areas that are on the 80,000 Hours Job Board but aren't at one of a handful of explicitly EA-motivated orgs" and "do earn to give for a while while gaining skills and then maybe transition to more direct work later or maybe not," as well as other paths that are specific to particular priority causes, e.g. for AI strategy & policy I'd be excited to see EAs (a) train up in ML, for later work in either AI safety or AI strategy/policy, (b) follow these paths into a US AI policy career (esp. for US citizens, and esp. now that CSET exists), and (c) train up as a cybersecurity expert (I hope to say more later about why this path should be especially exciting for AI-interested EAs; also the worst that happens is that you'll be in extremely high demand and highly paid).
A speculative thought I just had on one possible reason for why some people are overly focussed on EA jobs relative to e.g. the other options you list here:
- Identifying one's highest-impact career option is quite challenging, and there is no way to easily conclusively verify a candidate answer.
- Therefore (and for other reasons), many people rely a lot on advice provided by 80K and individual EAs they regard as suitable advisors.
- At least within the core of the (longtermist) EA community, almost all sources of advice agree that one of the most competitive jobs at an explicitly EA-motivated org usually is among the top options for people who are a good fit.
- However, for most alternatives there is significant disagreement among the most trusted sources of advice on whether these alternatives are competitive (in terms of expected impact) with an 'EA job', or indeed good ideas at all. For example, someone who I believe many people consult for career advice discouraged me from 'train up as a cybersecurity expert' - an option I had brought up (and according to my own impression still consider an attractive option) -, at least relative to working at an EA org. Similarl
... (read more)Two additional possible reasons:
- Many people in EA community believe it is easier to get a job at an EA organisation than it really is. People working at EA organisations, sometimes in senior positions, were surprised when they heard I didn't get an offer (from another organisation). I'd guess around half the organisations I applied to were "surprised about the very strong field of applicants". Past messaging about talent constraints probably also plays a role. As a result, career advice in the EA community can be overly optimistic, to a point where more than one person seriously encouraged me to apply for the COO position at OpenPhil (a position which went to the person who led the operations for Hillary Clinton's election campaign(!)). At least a year ago, when I was talking to dozens of people for career advice, I got the impression that it should be comparatively easy to get hired by an EA organisation.
- This one is weirdly specific and only a minor point (so this comment should not be misconstrued as "the two main reasons people apply for (too) many positions at EA organisations"). I don't know if this applies to many people, but I got qu
... (read more)Thanks for sharing your comment about personalized invitations, that's interesting. At Open Phil, almost all our personalized invitations (even to people we already knew well) were only lightly personalized. But perhaps a noticeable fraction of people misperceived that as "high chance you'll get the job if you apply," or something. The Open Phil RA hiring committee is discussing this issue now, so thanks for raising it.
[Deleted]
I have quantitative data on that :-)
I asked 10 people for career advice in a semi-structured way (I sent them the same document listing my options and asked them to provide rankings). These were all people I would think rank somewhere between "one of the top cause prioritization experts in the world" to "really, really knowledgeable about EA and very smart".
6 out of 10 thought that research analyst for OpenPhil would be my best option. But after that, there was much less consensus on the second best option (among my remaining three top options). 3.5 people rated management at an EA organisation highest, 3 rated biosecurity highest, 3.5 rated an MSc in ML (with the aim of doing AI safety research) highest.
Of course, YOU were one of these ten people, so that might explain some of it :-).
I had many more informal discussions, and I didn't think there was strong consensus either.
(Let me know if you need more data, I have many spreadsheets full of analysis waiting for you ;-) )
Sounds plausible. E.g. I'm pro "train up as a cybersecurity expert" but I know others have advised against.
In a nutshell, I'm worried that the people would not find the options you list exciting from their perspective, and instead would perceive not working in one of the 20 most competitive jobs at explicitly EA-motivated employers as some kind of personal shortcoming, hence the frustration.
I think the OP is evidence that his can happen e.g. because the author reports that
Note that I agree with you that in fact "[t]here are lots of exciting things for new EAs" including the options you've listed. However, even given this considered belief of mine, I think I was overly focussed on 'EA jobs' in a wa... (read more)
Thanks for this, we can clearly do better. Some ideas:
I don't really agree with your second and third point. Seeing this problem and responding by trying to create more 'capital letter EA jobs' strikes me as continuing to pursue a failing strategy.
What (in my opinion) the EA Community needs is to get away from this idea of channelling all committed people to a few organisations - the community is growing faster* than the organisations, and those numbers are unlikely to add up in the mid term.
Committing all our people to a few organisations seriously limits our impact in the long run. There are plenty of opportunities to have a large impact out there - we just need to appreciate them and pursue them. One thing I would like to see is stronger profession-specific networks in EA.
It's catastrophic that new and long-term EAs now consider their main EA activity to be to apply for the same few jobs instead of trying to increase their donations or investing in non-'capital letter EA' promising careers.
But this is hardly surprising given past messaging. The only reason EA organisations can get away with having very expensive hiring rounds for the applicants is because there are a lot of strongly committed people out there willing to take on that cost. Organisations cannot get away with this in most of the for-profit sector.
*Though this might be slowing down somewhat, perhaps because of this 'being an EA is applying unsuccessfully for the same few jobs' phenomena.
I disagree that "organizations cannot get away with this in most of the for-profit sector", at least when it comes to the kinds of for-profit jobs people in EA are likely to apply for.
I applied to ~10 different EA roles in 2018 (depending on how you count):
- The longest process, from Open Phil, involved roughly the same number of rounds, and the same amount of time, as the most time-intensive job I applied to out of college (at a hedge fund). They paid me for my time; the hedge fund didn't.
- CEA was a round shorter than that, and involved maybe 6 total hours of work before my work trial (at which point I had a very good chance of being hired -- also, I was paid at a reasonable rate during the trial).
- Of the other positions where I reached the final round or got an offer, none took more time than the job I accepted out of college (at a software company); most were along the lines of "one work test, a short interview, and a longer interview or set of interviews on-site". This seems to me like the standard in several high-skilled industries.
- Out of roughly 20 jobs and internships I applied for in college (and reached an interview round for), none of them took less
... (read more)When I applied to Google I did a phone interview and a full day of in-person interviews, plus a 1-hour conference call about how to do well in the second round. Lots of people devote significant time brushing up their coding interview skills as well; I only didn't because things like Project Euler had brushed up those skills for me.
Thanks for your comments. I already have a draft for a follow-up post on how I think the EA community could improve. Hopefully I will have time to write it up soon. Your points all seem to be good suggestions (with the caveat Denise mentions) .
More emotional data point that I wish I'd seen people expressing when I was new:
If you are excited about EA, relatively new, and eager to get involved in a way that looks something like "make this my career" rather than "make this a notable interest in my life," I think you should expect to feel a ton of significant pain, and reflect carefully on if that's worth it for you.
My experience trying to "break in" to the EA movement's professional class has at times made me feel like a worse person, less connected to my work, less confident in my ability to try my best, and basically more deeply miserable than any other professional situation I've been in by far. Other people I've met have described similar emotional experiences, or said their friends have had them.
If you've always had an easy/nice time at work, it can be hard to appreciate how bad this can feel up front.
I am not convinced that even people who succeed are particularly happy. I suspect it is happier on balance to have an ordinary job, read interesting things about altruism online, occasionally give talks at your local college, and give away 10% of your income to a... (read more)
Sorry to hear how much misery you've experienced. I'm curious to ask a follow-up question, but feel free to ignore if you aren't comfortable answering.
In particular, I'm wondering whether "make [EA] my career" feels ~identical (to you) to "work at a handful of explicitly EA-motivated employers." If it does, then maybe the messaging or energy or something in the EA community is pretty far from what I think it should be, which is more like what I said in another comment:
... (read more)The problem (for people like me, and may those who enjoy it keep doing so), as I see it: this is an elite community. Which is to say, this is a community primarily shaped by people who are and have always been extremely ambitious, who tend to have very strong pedigrees, and who are socialized with the norms of the global upper/top professional class.
"Hey you could go work for Google as a machine learning specialist" sounds to a person like me sort of like "Hey you could go be an astronaut." Sure, I guess it's possible. "Hey you could work for a nice nonprofit with all these people who share your weird values about charity, and join their social graph!" sounds doable. Which makes it a lot more damaging to fail.
People like me who are standardized-test-top-tier smart but whose backgrounds are pretty ordinary (I am inspired to post this in part because I had a conversation with someone else with the exact same experience, and realized this may be a pattern) don't tend to understand that they've traveled into a space of norms that is highly different than we're used to, when we join the EA social community. It just feels like "Oh! G... (read more)
Sorry to hear about your long, very difficult experience. I think part of what happened is that it did in fact get a lot harder to get a job at leading EA-motivated employers in the past couple years, but that wasn't clear to many EAs (including me, to some extent) until very recently, possibly as recently as this very post. So while it's good news that the EA community has grown such that these particular high-impact jobs can attract talent sufficient for them to be so competitive, it's unfortunate that this change wasn't clearer sooner, and posts like this one help with that, albeit not soon enough to help mitigate your own 1.5 years of suffering.
Also, the thing about some people not having runway is true and important, and is a major reason Open Phil pays people to take our remote work tests, and does quite a few things for people who do an in-person RA trial with us (e.g. salary, health benefits, moving costs, severance pay for those not made a subsequent offer). We don't want to miss out on great people just because they don't have enough runway/etc. to interact with our process.
FWIW, I found some of your comments about "elite culture" surprising. For context: I grew up in rur
... (read more)Thanks for this context, and for your warm replies in general. I really do feel okay now, I just also feel like I want people to not fall through the same sorts of cracks that I did in the future. I should also be clear that in my experience, lots of people in EA are from a wide variety of backgrounds. But enough core EAs are from a more elite background that's hard to detect up front, and the culture shock, at least for me, was massive.
I think EA culture is really good relative to general elite culture as I understand it. It's still the coolest professional culture I've been a part of. But I think norms really are a bit different than what I'm used to, in ways I find hard to place, and beyond the ways that are deliberate and reflectively good.
On further reflection re: enthusiasm, I think it's mostly a difference in enthusiasm around gratitude, specifically. People seem to display gratitude in a different way, which feels a lot more effusive.
I think the EA community is unusually meritocratic, though of course I agree it's imperfect. I'm glad people are working on making it even better. The fact that it's unusually meritocratic does make it a... (read more)
>"The problem (for people like me, and may those who enjoy it keep doing so), as I see it: this is an elite community. Which is to say, this is a community primarily shaped by people who are and have always been extremely ambitious, who tend to have very strong pedigrees, and who are socialized with the norms of the global upper/top professional class."
I wish this were shouted from the rooftops. Literally all the discourse around talent and jobs that I have come across to date in EA has frustrated me because of how this goes unremarked. As you say, many of the ideas that are discussed as the most natural and easy thing in the world are really like 'go be an astronaut' to normal humans. Having said that...
>"In elite culture, you're expected to be very positive in professional settings. You're expected to say "exciting" a lot, to call things "awesome," and to thank people creatively and effusively. In non-elite culture, there is no such expectation, and displays of extreme enthusiasm about work don't go over that well. Even at full enthusiasm-as-lived-experience you're unlikely to display it in the same way as so... (read more)
I really appreciate you writing this. You are not the first person to consider doing so and I applaud you for actually doing it.
People interested in this topic should read this post about how EAF ran their hiring round. They seem to have a good setup: only 25% of applicants were asked to take a work test, and only 25% of those tested were invited to a work trial (the other 64 of 68 were estimated to have invested <10 hours).
I'd love to see other organizations within the EA community publish similar numbers, though I understand that it can be tricky (e.g. someone might learn they were the only person interviewed who didn't make it to the next stage).
The following is a rough breakdown of the percentage of people who were not asked to move on to the next round in the Charity Science hiring process. These numbers assume one counterfactual hour of preparation for each interview and no preparation time outside of the given time limit for test tasks.
~3* hour invested (50%) - Cover letter/resume
~5 hours invested (20%) - Interview 1
~10 hours invested (15%) - Test task 1
~12 hours invested (5%) - Interview 2
~17 hours invested (5%) - Test task 2
~337 hours invested (2.5%) - paid 2-month work trial
Hired (2.5%)
So, 95% of those not hired spend 17 hours or less, 85% spend 14 hours or less, and 70% spend 5 hours or less.
*changed from 1 hour to 3 hours based on comments
1 hour for Cv/cover letter seems extremely optimistic...
For what it's worth, I think I'm unusually slow at this but I personally couldn't come close to writing a cover letter in an hour unless I had already written one for a nearly identical job (e.g. in college I applied for economics research assistants at most branches of the U.S. Federal Reserve; the cover letter for the Boston Fed and NY Fed were ~the same).
For a job I really cared about, I think three hours would be about the median if you don't count all the procrastination because I despise writing cover letters.
I'm *not* claiming this is typical. Just thought it would be useful to make it clear that there's a *really* wide range in how long job applications take applicants (perhaps depending primarily on level of neuroticism :P).
Ok given multiple people think this is off I have changed it to 3 hours to account for variation in application time.
It came from asking ~4 successful employees who where hired
Great post. I'm sure writing this must have been tough, so thanks very much for sharing this.
I wonder how much of the interview/work stuff is duplicated between positions - if there's a lot of overlap, then maybe it would be useful for someone to create the EA equivalent of TripleByte - run initial interviews/work projects with a third party organization to evaluate quality, pass along to most relevant EA jobs.
Thank you and respect for having written this. I really appreciate this, particularly you being open about this having been mentally challenging for you and the concrete data on time cost you provide.
Thanks for sharing this; strong upvoted! As someone running hiring rounds for EA orgs, this helped me better understand the perspective of applicants.
Same perspective here! Thank you for sharing.
I ran my first hiring process to hire someone for an EA role last year and was amazed how long it took me. I’ve hired around 20 times in the past and only spent a couple weeks and 20-40h per role. Last year I spent 8 months and hundreds of hours. I reflected afterwards on why and can list a few hypotheses:
normally rely heavily on gut to build my shortlist. Did not feel comfortable doing this for this role as it felt like there were so many failure modes for a bad hire. Both ways that a hire could go badly and severity of impact for a hire going badly.
normally relying on intuition heavily is highly reversible. Worst case scenario I have to fire the candidate after probation, I’ll never see them again and no one knows them. I’m open with candidates that this is my policy and that they should be careful accepting an offer. In EA I felt like everyone knows everyone and a fired hire could cause significant reputational damage with a one-sided narrative. I don’t endorse this view as rational but the fear was definitely a factor in why I took so long.
I was hiring for a role that defies regular role definition. No one applying to the role had applied to a similar position before l
"ultimately I made offers to two candidates both of which I had had strong gut feelings about very early, which was rewarding but also highly frustrating." - I hope this comment doesn't come across as incredibly mean, but, are you getting that from notes made at the time? When I find myself thinking "this is what I thought we'd do all along", I start to suspect I've conveniently rewritten my memories of what I thought. Do you have a sense of how many candidates you had similar strong positive gut feelings about?
Thank you for a very helpful comment!
I wonder whether this is just a result of people on both sides of the application process knowing each other in a social context.
If the candidate knows they will interact with people making the hiring decision in the future, they might not want them to feel bad about rejecting them. The people making the hiring decision might arguably feel less bad about not hiring someone if the candidate wasn't that excited. Lack of excitement also allows the candidate to save face if they get rejected, which also only matters because the candidate and the person making the hiring decision might interact socially in the future.
Meta note: I believe this is now the most upvoted EA forum post of all time by a wide margin. Seems like it struck a chord with a lot of people. It's probably worthwhile for people to write follow-up posts exploring issues related to human capital allocation, since it is a pretty central challenge for the movement. Example prompts:
A meta point: A lot of the discussion here has focused on reducing the time spent applying. I think a more fundamental and important problem, based on the replies here and my own experiences, is that many, many EAs feel that either they're working at a top EA org or they're not contributing much. Since only a fraction of EAs can currently work at a top EA org due to supply vastly exceeding demand, even if the time spent applying goes down a lot, many EAs will end up feeling negatively about themselves and/or EA when they get rejected. See e.g. this post by Scott Alexander on the message he feels he gets from the community. A couple of excerpts below:
... (read more)I agree that if it's true that "many EAs feel that either they're working at a top EA org or they're not contributing much," then that is much worse than anything about application time cost and urgently needs to be fixed. I've never felt that way about EA org work vs. alternatives, so I may have just missed that this is a message many people are getting.
E.g. Scott's post also says:
…and my reply is "Yes, talent-constrained also means those other things, and it's a big problem if that was unclear to a noticeable fraction of the community."
FWIW I suspect there's also something a bit more subtle going on than overly narrow misunderstandings of "talent-constrained," e.g. something like Max Daniel's hypothesis.
I think the votes for the old posts are not directly comparable with those in the new forum, since previously individuals could not give more than one upvote to a post. It may still be that this post would have been most upvoted of all time even if the new voting system would have been used for those old posts, however.
Good point, but this one has still received the most upvotes, if we assume that a negligible number of people downvoted it. At writing time, it has received 100 votes. According to https://ea.greaterwrong.com/archive, the only previous posts that received more than 100 points has less than 50 votes each. Insofar as I can tell, the second and third most voted-on posts are Empirical data on value drift at 75 and Effective altruism is a question at 68.
As an employer, I still think about this post three years after it was published, and I regularly hear it referenced in conversations about hiring in EA. The experiences in it clearly resonated with a lot of people, as evidenced by the number of comments and up votes. I think it's meaningfully influenced the frame of many hiring rounds at EA organizations over the past three years.
CEA run the EA community fund to provide financial support EA community group leaders.
The key metric that CEA for evaluating the success of the groups they fund is the number of people from each local group who reach interview stage for high impact jobs, which largely means jobs within EA organisations. Bonus points available if they get the job.
This information feels like a relevant piece of the puzzle for anyone thinking through these issues. It could be (that in hindsight) CEA pushing chapter organisers to push people to focus on jobs in EA organisations in many ways might not be the best strategy.
Hi, thanks for the comment.
I'm the project lead on EA Community Building Grants, which I think is what your referring to here.
The accredited positions include those that are in EA organisations, but many of the accredited positions are not. For example, accredited positions include doing an economics Phd at a top university, working at Facebook AI research, and many positions listed on the 80,000 Hours jobs board.
Based on the data submitted by groups so far, I'd expect the majority of positions we accredit to not be positions at EA organisations, though we're still in the early stages of receiving and analysing data here.
Our previous update post provides some more context on our evaluation criteria.
https://forum.effectivealtruism.org/posts/RZrikMAuTwt4e9Fs4/ea-community-building-grants-update
I'm sorry to see so many orgs take 10+ hours to get you only partway through the process, let alone multiple 40+ hour processes. This is especially glaring compared to the very low number of orgs that rejected you in under 5 hours.
It sounds like many of these orgs would benefit (both you and themselves!) from improving their evaluations to reject people earlier in the process.
My team at Wave's current technical interview process is under 10 hours over 4 stages (assuming you spend 1 hour on your cover letter and resume); the majority of rejections happen after less than 5 hours. The non-technical interview process is somewhat longer, but I would guess still not more than 15 hours and with the majority of applications being rejected in under 5 hours (the final interview is a full day).
Notably, we do two work samples, a 2hr one (where most applicants are rejected) and a 4-5hr one for the final interview. If I were interviewing for a non-technical role I'd insert a behavioral interview after the first work sample as well. These shorter interviews help us screen out many candidates before we waste a ton of their time. It's hard for me to imagine needing 8+ hours for a work sample unless the role is extremely complex and requires many different skills.
Wave is trying to do a much easier assessment than EA orgs mostly are; lots of people have thoughts about hire to software engineers, and software engineering is a well established industry with lots of established wisdom about how orgs should be structured. EA jobs often have much less precedent and so we shouldn't be surprised that we don't know how to figure out as efficiently whether people are likely to be good fits.
I think the reason the OP had a high fraction of 'long' processes had more to do with him being a strong applicant who would get through a lot of the early filters. I don't think a typical 'EA org' hiring round passes ~50% of its applicants to a work test.
This doesn't detract from your other points re. the length in absolute terms. (The descriptions from OP and others read uncomfortably reminiscent of more senior academic hiring, with lots of people getting burned competing for really attractive jobs). There may be some fundamental trade-offs (the standard argument about '*really* important to get the right person, so we want to spent a lot of time assessing plausible candidates to pick the right one, false negatives at intermediate stages cost more than false positives, etc. etc.'), but an easy improvement (mentioned elsewhere) is to communicate as best as one can the likelihood of success (perhaps broken down by stage) so applicants can make a better-informed decision.
Per Buck's comment, I think identifying software engineering talent is a pretty different problem than identifying e.g. someone who is already a good fit for Open Phil generalist RA roles.
Nearly all of Open Phil's RA hiring process is focused on assessing someone's immediate fit for the kind of work we do (via the remote work tests), not (other types of) fit with the team and mission.
I think there are probably a few things that some EA orgs could improve and I hope to write a post about it soon. In the meantime, it might be useful to explain where some of these high numbers come from:
1. Un-timed work test (e.g. OpenPhil research analyst):
I think most EA orgs underestimate how much time a work test takes. Take for example the conversation notes test of OpenPhil's application procedure. In the email instructions to the test, you will find the following line: "Historically, we think people have spent 2-8 hours on this assignment. " But there is no indication of how much time you should/are allowed to spend. And since everyone knows that the process is really competitive, and your results keep on improving if you invest more time, many people invest a lot of time. I spent 16 hours on the task. I asked three other people how much time they had spent, and it was 8 h, 16 h, and 24 h.
2. Research proposals (e.g. FHI research scholar programme, OpenPhil biosecurity early career researcher grant):
Writing a research proposal just takes a lot of time. I spent 30 hours on my proposal for FHI. I know of 4 other people who applied. These are the times they spen... (read more)
Thanks for mentioning the thing about the conversation notes test. It was simply an oversight to not explicitly say "Please don't spend more than X hours on this work test," and I've now added such a sentence to our latest draft of those work test instructions. We had explicit time limits for our other two tests.
I would advocate for controllably timed work tests whenever possible. Simply saying "please don't spend more than X hours on this work test" gives the opportunity to cheat by spending more time. Incentives for cheating are strong, because:
If you have enough applicants, some will cheat, and they will get a significant advantage. In rare cases, this may even deter people from applying. There was one position were I was planning to apply but then didn't because they had a non-controllably timed worktest (I don't want to cheat, somebody probably will cheat, and I am not super-well qualified for the position anyway so I would really need to shine in the work test -> not worth applying). (I admit that this deterrence probably doesn't happen often)
Great tools online for doing controllably timed work tests exist.
(I realize that it is not always possible to control the time limit, e.g. when the task is too long to be done... (read more)
Huh. I'm really surprised that they find this useful. One of the main ways that Wave employees' productivity has varied is in how quickly they can accomplish a task at a given level of quality, which varies by an order of magnitude between our best and worst candidates. (Or equivalently, how good of a job they can do in a fixed amount of time.) It seems like not time-boxing the work sample would make it much, much harder to make an apples-to-apples quality comparison between applicants, because slower applicants can spend more time to reach the same level of quality.
Two points that speak against this view a bit:
My current view is to ask for both timed and untimed tests, and make the untimed tests very simple/short (such that you could complete it in 20 minutes if you had to and there's very little benefit to spending >2h on it).
In case it isn't clear: EA is funding constrained. (I can't imagine it will ever not be. Have we ran out of imagination?) It is similarly difficult to raise funding as a budding EA organisation. The idea that EA is not funding constrained came from surveying established orgs that are indeed talent constrained, but in the meantime I could easily name 5 (even 10) startup EA charities that are perpetually seeking funding that they could instantly turn into jobs.
Even if that wasn't the case, higher pay would allow orgs to attract better talent, outsource more, etc. There's almost always a use for additional funding.
Especially if you compare the damage of too little with the damage of too much money, earning to give is still a great idea.
The fact that some organizations cannot get funded does not seem like strong evidence that EA as a whole is funding-constrained. Given that other organizations can raise large funds, an alternative explanation is that donors think that the expected impact of the organizations that cannot get funding is low. I also don't think it follows from your argument that earning to give is a great idea.
I agree that a failure to get funded does not imply funding constraints, but I definitely do think that many EA orgs, especially early ones, could benefit from more people with money looking to donate. There tends to be a large information asymmetry where you need to establish a clear track record and/or have someone spend a lot of time evaluating you before you can get funded. This is hard for early organizations to make happen.
I also think there are other systematic failures in EA where the best orgs do not always get fully funded.
That makes sense — on a second look, I misread your first comment. Absolutely agree that the community shouldn’t have a go big or go home mentality, ie it shouldn’t be seen as impossible to do good if you can’t get an ultra selective job at one of these organizations.
One quick recommendation for what applicants can do that might be useful for employers (speaking from an employer's perspective): Proactively share previous work trials. Whenever applicants did this, this provided me with additional valuable information that helped me decide whether someone should advance to the next stage.
I think you're correct, but my impression is that most EA orgs will happily agree to this.
Were you able to get good feedback as to why you weren't hired?
I got feedback for most of the roles where I reached the last stage of the application process, usually not for others. Usually the feedback was that my application was good (that I would have passed the hiring bar), but that someone else was better. I also got more specific feedback, but that was different every time (which makes some sense, since the positions were different, the people evaluating me were different, and I tried to not repeat mistakes that I had previously gotten feedback on).
Discussion between 80k & Peter Hurford in the comments here seems relevant.
Thank you for putting all of this together, I think it is a very useful post. I spent many years career coaching and advising people who were applying for jobs and I always stress this:
If you are not landing the job you want it is because of two main reasons:
a) You are not applying for the right jobs for you (you may be underqualified, overqualified, transitioning fields etc.)
b) Or you are in fact very well qualified but you are not good at presenting those qualifications to others, especially in a limited time and space.
Have you received any kind of feedback that would help you understand which was more common for you? If you think your case is a) then you need to find a different set of jobs to apply to - not better or worse, just different. If it is more b) then you may want to work on making sure that the reasons why you know you would be good at the job you are applying to translate into your resume, cover letter, interview etc.
I actually think it is quite common for people in the EA community to find themselves at a) because the jobs that are available to our community are very limited in numbers and scope. I think we need to expand the way we think about EA careers such that... (read more)
-
Your post is quite encouraging. I smiled while reading it when I came to this point "Yeah when we said that we need people, we meant capable people. Not you. You suck.” If people from educationally privileged countries like the US will find this difficult, what will be the fate of people like us who trained in Nigeria? Already you are educationally and positionally disadvantaged. I remember applying to 80000 hours for a one-on-one career conversation on how to make my career more impactful, I was surprised I got rejected.
I once heard the advice that if you are a donor who finds an opportunity that is worth EA money, you should just donate to it, even if you think that you could find a better opportunity after more research. Because if everyone only looks for the very best available opportunity, everyone will have to spend much more time evaluating many projects, or evaluations will be less deep.
The same advice could be adapted to hiring in some cases. If you find an ops person who is good enough to do ops for some EA org, you should consider hiring them, even if you think you could find a better candidate after more search. Because otherwise orgs will have to spend more time evaluating candidates, and candidates will have to spend more time applying.
In other words, having a lower threshold for hiring could be cooperative with other EA orgs in some game-theoretical scenario. Of course, if we go too far in this direction, we will no longer have a good grasp on where the threshold for being hired is, and best people might not get hired. And there are other complications. But EA orgs could go a little bit in this direction.
Having some experience with hiring, it might be some consolation that you did actually provide value to the EA orgs you applied to by giving them:
It's unfortunate that this has such a large opportunity cost and that you bore so much of it, but the unfortunate reality on the hiring side for any org is that we often need to interview additional candidates we know we won't hire because if we don't we won't know enough to trust the people we do hire are the right people. Of course we don't know which candidates we will and won't hire in advance (otherwise if we already knew it would be extremely unfair to everyone to do the interviews), but at least in this case your interview time helped EA focused organizations gain that knowledge rather than other orgs who you may be less aligned with values on and so less value their institutional learning from additional interviews that don't result in hires.
Thanks for sharing, and thanks to everyone for adding their own perspectives to this discussion. I would like to offer my view, which is informed by having gone through difficult periods myself.
I think its a mistake to get too involved with the view that getting a job in an EA org is the only way to make a difference in the world, or indeed to ourselves. In order to stay healthy I think we need to realize that there are other things that matter in the world than EA, such as our relations and our wellbeing. If we get too emotionally attached to the idea of making a huge, cosmic difference to the world by working at an EA org, we can get carried away and forget that we also have a responsibility towards ourselves. As an added benefit, if we take care of ourselvces we are happier, more resilient, can take on more risk, and more willing to work towards a long-term goal, all of which are good personality traits to eventually have higher impact in life.
Just wanted to give a quick note of encouragement that earning to give for long term future or EA meta causes can be very impactful. According to a survey of community leaders last year, donations to the long term future and EA meta EA funds were usually considered even more cost effective than donations to the animal welfare and global poverty funds.