Sometimes I wonder if people just don’t realize what very likely is in store in the future.

What life would be like after a nuclear war. What would constitute day to day existence when hundreds of millions of humans are desperately migrating to escape climate induced famine. Such things aren’t polite topics. They are, however, necessary topics.

The bulletin of the Atomic Scientists argument in favor of quantifying nuclear risk shares the following from Martin Hellman from Stanford University: 

"A risk of one percent per year would accumulate to worse-than-even odds over the lifetime of a child born today. Even if someone were to estimate that the lower bound should be 0.1 percent per year, that would be unacceptably high—that child would have an almost ten percent risk of experiencing nuclear devastation over his or her lifetime.'

The past half decade has seen a dramatic erosion in the taboo around threatening the use of nuclear weapons, even as they have proliferated across the globe. This threat will only be compounded and accelerated in the coming years as technological advance and proliferation continue to expand the risk of nuclear Armageddon. 

"Imagine that a man wearing a TNT vest were to sit down next to you and tell you that he wasn’t a suicide bomber. Rather, there are two buttons for setting off his explosive vest. One was in the White House with Trump for the last four years, and recently was given to Biden. The other is with Putin in Moscow. You’d still get away as fast as you can! " - Vinton Cerf

Yet when I look at Open Philanthropy's grants for nuclear deterrence, all I see are a few tangentially related longtermist grants on biosecurity. The term nuclear seems to just come in where organizations like Nuclear Threat Initiative work on biosecurity grants. Why is this the case? What is the logic there? This seems like an obvious blind spot.

I wonder if EA is subject to the same myopia and complacency that's gripped our civilization since the end of the Cold War. The nuclear threat feels dated and tired. Yet a clear eyed analysis shows that the threat has only grown. 

The war in Ukraine may very well be the start of World War Three. Do any of us have the right to be surprised? What then is the EA response to this threat? Worship at the cult of the coming AGI singularity and escape into cyberspace? Pat ourselves on the backs for the number of bednets we've funded? I'm being ungenerous in my rhetoric but what seriously are we doing people? 

What is preventing us from showing much greater leadership and courage in the face of what may very well be the end of humanity? Refuges and bunkers are great. Why not more? Why not take a more proactive role? Why not work to address the underlying threat at its source? 

1

0
0

Reactions

0
0
Comments5
Sorted by Click to highlight new comments since: Today at 7:18 PM

I think this is caused by two factors, 1) longtermism has decided to distinguish between existential risks and global catastrophic risks, and 2) there is a much stronger general culture of denial around nuclear risk than any other risk. 

When in the Precipice(which I take to have been broadly read) Toby Ord says that AI is a 1/10 risk and nuclear war and climate change are both 1/1000 and does so essentially on the basis that he views the later two as survivable this created the justification to write them off.(this isn't to say this is where the trend began)

If you talk to most people they just don't believe nuclear war is likely, as a result, it seems like nuclear war and unaligned artificial intelligence are both treated as being speculative, except nuclear war has a long history of not happening. This is of course nonsense, as nuclear weapons exist right now.

Yes 1000% on the cultural factors that have desensitized us to nuclear risk. Tyler Cowen has a nice series of posts out today on this subject: https://marginalrevolution.com/marginalrevolution/2022/08/which-is-the-hingy-est-century.html

I might find time to go back and reread the Precipice and dig into the probabilities you reference. Those seem odd. It's also odd because something that reduces humanity to subsistence levels for a very long time and eliminates ninety some odd percent of the population is absolutely catastrophic. I suppose I'm a hyberbolic discounter at heart and do think that while we should care about the far future, it's really silly to get into the 1-1 logic that a human a billion years from now should be equally valued for decision making as one today or ten years from now. 

I'll check out the article. 

You can find the numbers in the Risk Landscape section. Yeah, it's wonky. It seems very odd to me to both A) be confident that nuclear risk won't kill everyone but will just kill 90% of people and B) that this would strongly influence which risks you worry about. 

Also, you might be interested in the submission I just made to the contest about many worlds and nuclear risk, as I am also begging EA to care about nuclear risk. https://forum.effectivealtruism.org/posts/Gg2YsjGe3oahw2kxE/nuclear-fine-tuning-how-many-worlds-have-been-destroyed?

Thanks for sharing. I'll check out your post. 

Do you know how one might get another copy of the Precipice? I donated mine to a friend.

If you sign up for the 80,000 hours newsletter you can get one for free, it is also on libgen

Curated and popular this week
Relevant opportunities