All of Ryan Beck's Comments + Replies

One thing worth noting is that extinction and extreme economic collapse were excluded from all except 2 questions (GDP and total population) to make the forecasts more interpretable (more info on this in the appendix). This might explain some of the confidence you see, though it also might not!

This comment was copied from a reply I made on the EA forecasting and epistemics slack.

What's the basis for using expected utility/value calculations when allocating EA funding for "one off" bets? More details explaining what I don't understand are below for context.

My understanding is expected value relies on the law of large numbers, so in situations where you have bets that are unlikely to be repeated (for example, extinction, where you could put a ton of resources into it and go from a 5% extinction risk over the next century to a 4% risk) it doesn't seem like expected value should hold. The way I've seen this justified is using expected... (read more)

2
Howie_Lempel
2y
It's been a while since I read it but Joe Carlsmith's series on expected utility might help some. 

My two cents on a couple of these from the perspective of a father of two girls (4 years old and 2 years old, I'm 30). Just my perspective, feel free to disregard if not helpful!

On emotions and discipline, I'm also a very calm person and rarely show anger and frustration. But kids are really good at finding was to frustrate you. I almost never yell at them, but I do get frustrated or exasperated and raise my voice, and it's genuinely unclear to me how anyone could parent a child without doing that.

In general I think the military wisdom "no plan survives fi... (read more)

I always interpreted the 10% as a goal, not a requirement for EA. That's a pretty high portion for a lot of people. I worry that making that sound like a cutoff makes EA seem even more inaccessible.

The way I had interpreted the community message was more like "an EA is someone that thinks about where their giving would be most effective or spends time working on the world's most pressing problems."

Thanks for writing this, as a new-ish user of the forum it's been frustrating trying to find previous posts that address questions I have or things I want to learn more about, only to find sprawling or multi-part posts with half hour or longer read-times that may or may not address the specific thing I'm interested in.

Also you mentioned jargon and I think there's room for a lot of improvement there, it seems to me like there's more jargon than is justified and it made the forum daunting for me. This previous post has some good recommendations and in my opi... (read more)

Great post as usual.

It looks like your Putin's health link goes to the wrong forecast.

3
NunoSempere
2y
Thanks, fixed.

I've found this short article useful in explaining the case for it. Basically it says that a guarantee of defense could embolden Taiwan to more aggressively pursue independence which could provoke China, while committing to not interfere could embolden China to invade. The US benefits from better relations with both countries if it walks a line between them and it may be better for peace between them if Taiwan has to tread carefully and China expects a high chance of the US fighting off an invasion of Taiwan.

Thanks for posting this, I'm glad to see more discussion of the issue and you've laid it out very nicely.

In the interest of thinking seriously about this potential deadly conflict, could you explain why you lean toward abandoning Taiwanese independence if war appears likely? Aside from principle based stances about protecting potential allies and the right of countries to continue governing themselves, I think my main worry is that giving in to bullying seems like it would incentivize future bullying. If the US and other nations declare that they no longer... (read more)

7
JamesHu
2y
Let's limit our consideration to upholding international norms and laws of non-aggression, which I think is the crux of your argument for a more hardline stance on Taiwan. The sentiment of wanting to uphold these norms and laws is admirable, and in an ideal world where they are strong we should be willing to expend significant resources and even risk nuclear conflict to keep those norms that way. But many would argue those norms are already in tatters, in large part due to the US's repeated and flagrant violations thereof, and that a hardline Western response to Russian/Chinese aggression will have minimal benefit for preserving whatever baseline of norms we still have.  (And that is even leaving out that Taiwan has the further complication of being, unlike Ukraine, an unrecognized state, which further dilutes the argument of international norms and laws.) This tilts the cost-benefit analysis significantly toward a less hardline stance. I'm taking this from one-time 80,000 Hours Podcast guest Robert Wright's Nonzero Newsletter – in my view, one of the best Substacks out there (I also highly recommend the Wright Show podcast) – which recently had a post calling for peace talks in Ukraine that lays out this line of argument: At the same time, we should seriously consider the leftist critique of neoliberal foreign policy, which Hfur7c was perhaps, inartfully, trying to espouse: that proactive diplomatic efforts (or even just basic, responsive efforts, which many argue the US did not engage seriously in Ukraine) have great potential to forestall conflict, and that strong military stances in themselves can provoke conflict, an outcome that is then used to justify dispensing with robust diplomatic efforts. See this Twitter thread from the Marxist historian Jake Werner:
-30
Hfur7c
2y

Your comment is 3 months old, but somehow I missed it back when I was posted and am just now seeing it, so I just wanted to say these are all good points, particularly about cooperation on other issues like your climate example!

It's possible I missed it but I didn't see anything stating whether multiple submissions from one author are allowed, I assume they are though?

3
Gavin
2y
Don't see why not, as long as it's not salami sliced.

Is there a way to sort answers by newest? I'm not seeing that option. It would be useful for finding new answers I haven't seen yet.

It's now possible to sort answers by date! \o.o/

4
Lorenzo Buonanno
2y
I don't know if there's an easy way to sort by newest, made a quick and dirty codepen using the forum API https://codepen.io/lorenzo-buonanno/full/xxYjvgW 
Answer by Ryan BeckMay 27, 202216
0
0

Update July 19, 2022 - I've accepted an employment offer and am currently not seeking a new role

Location: Council Bluffs, Iowa

Remote: Yes

Willing to relocate: No

Skills: 

... (read more)

I'm not sure how your first point relates to what I was saying in this post; but, I'll take a guess.

Sorry, what I said wasn't very clear. Attempting to rephrase, I was thinking more along the lines of what the possible future for AI might look like if there were no EA interventions in the AI space. I haven't seen much discussion of the possible downsides there (for example slowing down AI research by prioritizing alignment resulting in delays in AI advancement and delays in good things brought about by AI advancement). But this was a less-than-half-bake... (read more)

Did your outcomes 2 and 3 get mixed up at some point? I feel like the evaluations don't align with the initial descriptions of those, but maybe I'm misunderstanding.

Thanks for writing this though, this is something I've been thinking a little about as I try to understand longtermism better. It makes sense to be risk-averse with existential risk, but at the same I have a hard time understanding some of the more extreme takes. My wild guess would be that AI has a significantly higher chance of improving the well-being of humanity than it does causing extinct... (read more)

1
Joshc
2y
Yep, thanks for pointing that out! Fixed it. I'm not sure how your first point relates to what I was saying in this post; but, I'll take a guess. I said something about how investing in capabilities at anthropic could be good. An upside to this would be increasing the probability that EAs end up controlling the super-intelligent AGI in the future. The downside is that it could shorten timelines, but hopefully this can be mitigated by keeping all of the research under wraps (which is what they are doing). This is a controversial issue though. I haven't thought very much about whether the upsides outweigh the downsides, but the argument in this post caused me to believe the upsides were larger than I thought before.   It doesn't matter what outcome you assign zero value to as long as the relative values are the same since if a utility function is an affine function of another utility function then they produce equivalent decisions.

It's a common misconception that those who want to mitigate AI risk think there's a high chance AI wipes out humanity this century. But opinions vary and proponents of mitigating AI risk may still think the likelihood is low. Crowd forecasts have placed the probability of a catastrophe caused by AI as around 5% this century, and extinction caused by AI as around 2.5% this century. But even these low probabilities are worth trying to reduce when what's at stake is millions or billions of lives. How willing would you be to take a pill at random from a pile o... (read more)

That's a good point, I agree. None of my suggestions really fit very well, it's hard to think of a descriptive name that could be easily used conversationally.

1
Ben Stewart
2y
They're good attempts though - I think this is just a tricky needle to thread

It still seems like prefixing with "not" still runs into defining based on disagreement, where I would guess people who lean that way would rather be named for what they're prioritizing as opposed to what they aren't. I came up with a few (probably bad) ideas along that vein:

  • Immediatists (apparently not a made up word according to Merriam-Webster)
  • Contemporary altruists
  • Effective immediately

I'm relatively new so take my opinion with a big grain of salt. Maybe "not longtermist" is fine with most.

2
Ben Stewart
2y
To me 'contemporary altruists' suggests people who are alive today and altruistic, in contradistinction to historical altruists in the past, e.g. Katharine McCormick or John D. MacArthur. 
4
david_reinstein
2y
I see your point but I don't think the non-Ltists/neartermists actually do identify as a group along those lines (I may be wrong here). So for me, just "non-LTist EA" seems the right descriptor. Although "Global Health and Wellbeing" (or maybe just "Global Wellbeing") seem pretty decent. I could see "Non-LTist EA" as the term to use for precision, and then also identify people by the cause, approach, or moral philosophy they care most about.

There are good points and helpful, thanks! I agree I wasn't clear about viewing the scenarios exclusively in the initial comment, I think I made that a little clearer in the follow up.

when I read 80% to reach saturation at 40% predictive power I read this as "capping out at around 40%" which would only leave a maximum of 20% for scenarios with much greater than 40%?

Ah I think I see how that's confusing. My use of the term saturation probably confuses things too much. My understanding is saturation is the likely maximum that could be explained with curr... (read more)

No problem!

Also if you're interested in elaborating about why my scenarios were unintuitive I'd appreciate the feedback, but if not no worries!

2
elifland
2y
At first I thought the scenarios were separate so they would be combined with an OR to get an overall probability, which then made me confused when you looked at only scenario 1 for determining your probability for technological feasibility. I was also confused about why you assigned 30% to polygenic scores reaching 80% predictive power in Scenario 2 while assigning 80% to reaching saturation at 40% predictive power in the Scenario 1, because when I read 80% to reach saturation at 40% predictive power I read this as "capping out at around 40%" which would only leave a maximum of 20% for scenarios with much greater than 40%? Finally, I was a little confused about where the likelihood of iterated embryo selection fit into your scenarios; this seems highly relevant/important and is maybe implicitly accounted for in e.g. "Must be able to generate 100 embryos to select from"? But could be good to make more explicit.

This was a cool contest, thanks for running it! In my view there's a lot of value in doing this. Doing a deep dive into polygenic selection for IQ was something I had wanted to do for quite a while and your contest motivated me to finally sit down and actually do it and to write it up in a way that would be potentially useful to others.

I think your initial criteria of how much a writeup changed your minds may have played a role in fewer than expected entries as well. Your forecasts on the set of questions seemed very reasonable and my own forecasts were pr... (read more)

2
elifland
2y
Thanks for sharing Ryan, and that makes sense in terms of another unintended consequence of our judging criteria; good to know for future contests.
3
Misha_Yagudin
2y
Yes, thank you; that makes sense and is very helpful!

That's really interesting, thanks! I wonder why India is so supportive of it in comparison to other countries.

9
Question Mark
2y
A major reason why support for eugenically raising IQs through gene editing is low in Western countries could be a backlash against Nazism, since Nazism is associated with eugenics in the mind of the average person. The low level of support in East Asia is more uncertain. One possible explanation is that East Asians have a risk-averse culture. Interestingly, Hindus and Buddhists also have some of the highest rates of support for evolution among any religious groups. There was a poll from 2009 that showed that 80% of Hindus and 81% of Buddhists in the United States accept evolution, while only 48% of the total US population accepts evolution. Another poll showed that 77% of Indians believe that there is significant evidence to support evolution. The high rate of acceptance of gene editing technology among Hindu Indians could therefore be a reflection of greater acceptance of science in general.

Even if the ~300 new DF-41 silos discovered last year are each armed with only 3 warheads (the missile can carry ~10 max), and no other silos are built/discovered, that's still 900 warheads on top of the ~400 already in service.

I'm not well-versed in this area but reading through the Chinese nuclear notebook from November 2021 they seem kind of skeptical of claims like this and point out that China could also be intending the silos to be a "shell game". Quoting from the notebook:

And in November 2021, the Pentagon’s annual report to Congress projects t

... (read more)

Yeah definitely on the same page then! I agree with what you said there with the possible exception or caveat that I'm skeptical on improvements to the Taiwan issue and that if you find or know of any persuasive abyss-staring arguments on this topic (or write them yourself) I'd appreciate it if you share them with me because I'd be happy to be wrong in my skepticism and would like to learn more about any promising options.

To be clear I'm not arguing that people shouldn't think about it or try to solve it. I'm definitely in favor of more discussion on that topic and I'd love to read some high effort analysis from an EA perspective.

If I'm understanding correctly the main point you're making is that I probably shouldn't have said this:

There is little room for improvement here...

Which in that case that's a fair critique. I'm not well-informed enough to know the options here and their advantages and risks in great detail, so my perception that there's not much room for impro... (read more)

2
Matthew_Barnett
2y
Thanks for the continued discussion. I think I'm making two points. The first point was, yeah, I think there is substantial room for improvement here. But the second point is necessary: analyzing the situation with Taiwan is crucial if we seek to effectively reduce nuclear risk. I do not think it was wrong to focus on the trade war. It depends on your goals. If you wanted to promote quick, actionable and robust advice, it made sense. If you wanted to stare straight into the abyss, and solve the problem directly, it made a little less sense. Sometimes the first thing is what we need. But, as I'm glad to hear, you seem to agree with me that we also sometimes need to do the second thing.

This is a good point, I completely agree that the trade war is of small importance relative to things like relations with Taiwan. My reason for focusing on the trade war though is because trade deescalation would have very few downsides and would probably be a substantial positive all on its own before even considering the potential positive effects it could have on relations with China and possibly nuclear risk.

To me the same can't be said for the Taiwan issue. The optimal policy here is far from clear to me. Strategic ambiguity is our intentional policy,... (read more)

2
Matthew_Barnett
2y
I agree. I think we're both on the same page about the merits of ending the trade war, as an issue by itself. Right. From my perspective, this is what makes focusing on Taiwan precisely right thing to do in a high-level analysis. My understanding of your point here is something like, "The US-Taiwan policy is a super complicated issue so I decided not to even touch it." But, since the US-Taiwan policy is also the most important question regarding US-China relations, not talking about it is basically just avoiding the hard part of the issue.  It's going to be difficult to make any progress if we don't do the hard work of actually addressing the central problem. (Maybe this is an unfair analogy, but I find what you're saying to be a bit similar to, "I have an essay due in 12 hours. It's on an extremely fraught topic, and I'm unsure whether my thesis is sound, or whether the supporting arguments make any sense. So, rather than deeply reconsider the points I make in my essay, I'll just focus on making sure the essay has the right formatting instead." I can sympathize with this sort of procrastination emotionally, but the clock is still ticking.) I expect experts to have basically spilled a huge amount of ink about every policy regarding US-China relations, so I don't see this as a uniquely asymmetric argument against thinking about Taiwan. Maybe your point is merely that these experts have not yet come to a conclusion, so it seems unlikely that you could come to a conclusion in the span of a short essay. This would be fair reply, but I have two brief heuristic thoughts on that, 1. Most international relations experts neither understand, nor are motivated by an EA mindset. To the extent that you buy EA philosophy, I think we are well-positioned to have interesting analyses on questions such as, "Is it worth risking nuclear war to save a vibrant democracy?" It's not clear to me at all that moral philosophers have adequately responded to this question already, in the

That's a good question, I've thought about this some before and while it's kind of messy I think the general gist of my thoughts is something like this (framed from a US perspective but I think it generalizes to most countries):

Tariffs should be avoided or minimized wherever possible due to them likely costing US citizens much more than they benefit them. However tariffs and sanctions can be important tools when a country does something very offensive, particularly when punitive measures are applied in cooperation with allies. Tariffs and sanctions shoul

... (read more)
3
Stefan_Schubert
2y
Thanks for the response. Yeah I agree that starting a trade war over trade imbalances is significantly different from initiating human rights-motivated sanctions. 

That makes sense, I agree it's better to have more direct sources.

I don't know this area at all, but here is data from one review paper I found.

I didn't have access to your link but I found another version of it here.

To be honest I'm not familiar with the direct evidence either so I'm mostly relying on secondhand impressions and general descriptions of tariff burdens falling on consumers. I searched around briefly just now and found this paper (also cited in the paper you linked as Amiti et al. (2020b)) which reports:

Using another year of data including significant escalations in the trade war, we find that U.S. tariffs

... (read more)
3
Stefan_Schubert
2y
Thanks. Tbc I don't have a strong view on the object-level issue. I just thought it would be good to have more evidence than that early survey.