Hi Remmelt,
Thanks for sharing your concerns, both with us privately and here on the forum. These are tricky issues and we expect people to disagree about how to about how to weigh all the considerations — so it’s really good to have open conversations about them.
Ultimately, we disagree with you that it's net harmful to do technical safety research at AGI labs. In fact, we think it can be the best career step for some of our readers to work in labs, even in non-safety roles. That’s the core reason why we list these roles on our job board.
We argue for this p...
Most of our advice on actually having an impact — rather than building career capital — is highly relevant to mid-career professionals. That's because they're entering their third career stage (https://80000hours.org/career-guide/career-planning/#three-career-stages), i.e. actually trying to have an impact. When you’re mid-career, it's much more important to appropriately:
So we hope mid-career people can get a lot out of reading our articles. I'd probably i...
Thanks for this comment Tyler!
To clarify what I mean by unknown unknowns, here's a climate-related example: We're uncertain about the strength of various feedback loops, like how much warming could be produced by cloud feedbacks. We'd then classify "cloud feedbacks" as a known unknown. But we're also uncertain about whether there are feedback loops we haven't identified. Since we don't know what these might be, these loops are unknown unknowns. As you say, the known feedback loops don't seem likely to warm earth enough to cause a complete destruction of ci...
I don't currently have a confident view on this beyond "We’re really not sure. It seems like OpenAI, Google DeepMind, and Anthropic are currently taking existential risk more seriously than other labs."
But I agree that if we could reach a confident position here (or even just a confident list of considerations), that would be useful for people — so thanks, this is a helpful suggestion!
Thanks, this is an interesting heuristic, but I think I don't find it as valuable as you do.
First, while I do think it'd probably be harmful in expectation to work at leading oil companies / at the Manhattan project, I'm not confident in that view — I just haven't thought about this very much.
Second, I think that AI labs are in a pretty different reference class from oil companies and the development of nuclear weapons.
Why? Roughly:
Hi Yonatan,
I think that for many people (but not everyone) and for many roles they might work in (but not all roles), this is a reasonable plan.
Most importantly, I think it's true that working at a top AI lab as an engineer is one of the best ways to build technical skills (see the section above on "it's often excellent career capital").
I'm more sceptical about the ability to push towards safe decisions (see the section above on "you may be able to help labs reduce risks").
The right answer here depends a lot on the specific role. I think it's importa...
The Portuguese version at 80000horas.com.br is a project of Altruísmo Eficaz Brasil. We often give people permission to translate our content when they ask - but as to when, that would be up to Altruísmo Eficaz Brasil! Sorry I can't give you a more concrete answer.
(Personal views, not representing 80k)
My basic answer is "yes".
Longer version:
I think this depends what you mean.
By "longtermism", I mean the idea that improving the long-run future is a key moral priority. By "longtermist" I mean someone who personally identifies with belief in longtermism.
I think x-risks are the most pressing problems from a cause-neutral perspective (although I'm not confident about this, there are a number of plausible alternatives, including factory farming).
I think longtermism is also (approximately) true from a cause neutral perspec...
Thank you so much for this feedback! I’m sorry to hear our messaging has been discouraging. I want to be very clear that I think it’s harmful to discourage people from working on such important issues, and would like to minimise the extent to which we do that.
I wrote the newsletter you’re referencing, so I particularly wanted to reply to this. I also wrote the 80,000 Hours article on climate change, explaining our view that it’s less pressing than our highest priority areas.
I don’t consider myself fundamentally a longtermist. Instead, I try my b...
There are important reasons to think that the change by the EA community is within the measurement error of these surveys, which makes this less noteworthy.
(Like say you put +/- 10 years and +/- 10% on all these answers - note there are loads of reasons why you wouldn't actually assess the uncertainty like this, (e.g. probabilities can't go below 0 or above 1), but just to get a feel for the uncertainty this helps. Well, then you get something like:
Then many many EA timelines an...
Thanks for this thoughtful post! I think I stand by my 1 in 10,000 estimate despite this.
A few short reasons:
Hi! Wanted to follow up as the author of the 80k software engineering career review, as I don't think this gives an accurate impression. A few things to say:
This looks really cool, thanks Tom!
I haven't read the report in full (just the short summary) - but I have some initial scepticism, and I'd love to answers to some of the following questions, so I can figure out how much evidence this report is on takeoff speeds. I've put the questions roughly in order of subjective importance to my ability to update:
Thanks for these great questions Ben!
To take them point by point:
I agree with (a). I disagree that (b) is true! And as a result I disagree that existing CEAs give you an accurate signpost.
Why is (b) untrue? Well, we do have some information about the future, so it seems extremely unlikely that you won't be able to have any indication as to the sign of your actions, if you do (a) reasonably well.
Again, I don't purely mean this from an extreme longtermist perspective (although I would certainly be interested in longtermist analyses given my personal ethics). For example, simply thinking about population changes in the abo...
Sure, happy to chat about this!
Roughly I think that you are currently not really calculating cost-effectiveness. That is, whether you're giving out malaria nets or preventing nuclear war, almost all of the effects of your actions will be affecting people in the future.
To clarify, by "future" I don't necessarily mean "long run future". Where you put that bar is a fascinating question. But focusing on current lives lost seems to approximately ignore most of the (positive or negative) value, so I expect your estimates to not be capturing much about what matte...
I'm curious about the ethical decisions you've made in this report. What's your justification for evaluating current lives lost? I'd be far more interested in cause-X research that considers a variety of worldviews, e.g. a number of different ways of evaluating the medium or long-term consequences of interventions.
I agree that I'd love to see more work on this! (And I agree that the last story I talk about, of a very fast takeoff AI system with particularly advanced capabilities, seems unlikely to me - although others disagree, and think this "worst case" is also the most likely outcome.)
It's worth noting again though that any particular story is unlikely to be correct. We're trying to forecast the future, and good ways of forecasting should feel uncertain at the end, because we don't know what the future will hold. Also, good work on this will (in my opinion) give ...
That particular story, in which I write "one day, every single person in the world suddenly dies", is about a fast takeoff self-improvement scenario. In such scenarios, a sudden takeover is exactly what we should expect to occur, and the intermediate steps set out by Holden and others don't apply to such scenarios. Any guessing about what sort of advanced technology would do this necessarily makes the scenario less likely, and I think such guesses (e.g. "hypnodrones") are extremely likely to be false and aren't useful or informative.
For what it's worth, I ...
Yeah, it’s a good question! Some thoughts:
I’m being quite strict with my definitions. I’m only counting people working directly on AI safety. So, for example, I wouldn’t count the time I spent writing this profile on AI (or anyone else who works at 80k for that matter). (Note: I do think lots of relevant work is done by people who don’t directly work on it) I’m also not counting people who think of themselves as on an AI safety career path and are, at the moment, skilling up rather than working directly on the problem. There are some ambiguities, e.g. a
Hi Gideon,
I wrote the 80,000 Hours problem profile on climate change. Thank you so much for this feedback! I’m genuinely really grateful to see such engagement with the things I write - and criticism is always a welcome contribution to making sure that I’m saying the right things.
Just to be clear, when I said “we think it’s potentially harmful to do work that could advance solar geoengineering”, I meant that (with a fair degree of uncertainty), it could be harmful to do work that advances the technology (which I think you agree with) not that all research ...
I think these are all great points! We should definitely worry about negative effects of work intended to do good.
That said here are two other places where maybe we have differing intuitions:
This is a great story! Good motivational content.
But I do think, in general, a mindset of "only I can do this" is innacurate and has costs. There are plenty of other people in the world, and other communities in the world, attempting to do good, and often succeeding. I think EAs have been a small fraction of the success in reducing global poverty over the last few decades, for example.
Here are a few plausible costs to me:
Knowing when and why others will do things significantly changes estimates of the marginal value of acting. For example, if you are st
I really like these nuances. I think one of the problems with the drowning child parable / early EA thinking more generally was (and still is, to a large extent) very focused on the actions of the individual.
It's definitely easier and more accurate to model individual behavior, but I think we (as a community) could do more to improve our models of group behavior even though it's more difficult and costly to do so.
This does seem to be an important dynamic.
Here are a few reasons this might be wrong (both sound vaguely plausible to me):
(1) is particularly important if you think this "non-weird to weird" ap...
I agree with both of those reasons in the abstract, and I definitely do (2) myself. I'd guess there are around 50 people total in the world who could do (2) in a way where I'd look at it and say that they succeeded (for AI risk in particular), of which I could name maybe 20 in advance. I would certainly not be telling a random EA to make our arguments sound less weird.
(EDIT: My estimate is now more like 30; I actually asked some people to do (2) for AI alignment and they did worse than I expected.)
I'd be happy about the version of (1) where the non-w...
That's not the intention, thanks for pointing this out!
To clarify, by "route", I mean gaining experience in this space through working on engineering roles directly related to AI. Where those roles are not specifically working on safety, it's important to try to consider any downside risk that could result from advancing general AI capabilities (this in general will vary a lot across roles and can be very difficult to estimate).
A bit of both - but you're right, I primarily meant "secure" (as I expect this is where engineers have something specific to contribute).
Totally agree! Indeed, there's a classic 80k article about this.
When working out your next steps, we tend to recommend working forwards from what you know, and working backwards from where you might want to end up (see our article on finding your next career steps). We also think people should explore more with their careers (see our article on career exploration).
If there are areas where we're giving the opposite message, I'd love to know – shoot me an email or DM?