Do people's wildness intuitions change when we think about human lives or life-years, instead of calendar years?
7 billion of 115 billion humans ever are living today. Given today's higher life expectancies, about 15% of all experience so far has been experienced by people who are alive right now.
So the idea that the reference class "humans alive today" is special among pre-expansion humans doesn't feel that crazy. (There's a related variant of the doomsday argument -- if we expect population to grow rapidly until space expansion, many generations might rea...
Nice to see the assumptions listed out. My worries about the future turning out well by default are part of why I'd like to see more work done in clarifying and sharing our values, and more work on questioning this assumption (eg looking into Pinker's work, thinking about why the trends might not hold). I'm aware of SI and some negative-utilitarian approaches to this, but I'd love to see links on whatever else is out there.
I think most EAs share these premises with you:
1. Some people live in relative material abundance, and face significant diminishing returns to having more material wealth.
2. However, many problems remain, including poverty and catastrophic risk.
3. It would be valuable for funds to go towards reducing these problems, and thus quite valuable to successfully spread values that promote donating towards them.
You also make a couple of interesting claims:
4. We can feasibly cause a 'paradigm shift' in values by convincing people to tithe.
5. The benefit...
Interesting talk. I agree with the core model of great power conflict being a significant catastrophic risk, including via leading to nuclear war. I also agree that emerging tech is a risk factor, and emerging tech governance a potential cause area, albeit one with uncertain tractability.
I would have guessed AI and bioweapons were far more dangerous than space mining and gene editing in particular; I'd have guessed those two were many decades off from having a significant effect, and preventing China from gene editing seems low-tractability. Geoengin...
This seems like a good point, and I was surprised this hadn't been addressed much before. Digging through the forum archives, there are some relevant discussions from the past year:
To provide more information on the status of the EA Angel Group, Benjamin Pence and I are working together on the EA Angel Group (and its parent project Altruism.vc). The EA Angel Group is operating, although it received a lower than expected number of referrals from angels within the group which has significantly reduced the benefit that the group currently provides to its members.
I anticipated this concern months ago and tried to resolve the issue, but was delayed by ~5 months in our attempt to discuss sharing grant proposals with EA Grants. I felt like ...
Thanks for the thoughts, Michael. Sorry for the minor thread necro - Milan just linked me to this comment from my short post on short-termism.
The first point feels like a crux here.
On the second, the obvious counterargument is that it applies just as well to e.g. murder; in the case where the person is killed, "there is no sensible comparison to be made" between their status and that in the case where they are alive.
You could still be against killing for other reasons, like effects on friends of the victim, but I think most people have an intui...
One terminology for this is introduced in "Governing Boring Apocalypses", a recent x-risk paper. They call direct bad things like nuclear war an "existential harm", but note that two other key ingredients are necessary for existential risk: existential vulnerability (reasons we are vulnerable to a harm) and existential exposure (ways those vulnerabilities get exposed). I don't fully understand the vulnerability/exposure split, but I think e.g. nuclear posturing, decentralized nuclear command structures, and launch-on-warning system...
You mention nanotechnology; in a similar vein, understanding molecular biology could help deal with biotech x-risks. Knowing more about plausible levels of manufacture/detection could help us understand the strategic balance better, and there’s obviously also concrete work to be done in building eg better sensors.
On the more biochemical end, there’s of mechanical and biological engineering for cultured meat.
Also, wrt non-physics careers, a major one is quantitative trading (eg at Jane Street), which seems to benefit from a physics-y mindset and use some similar tools. I think there’s even a finance firm that mostly hires physics PhDs.
Interesting, scary stuff. I've been reading up on biotech/bioweapons a bit as part of my research on AI strategy. They're interesting both because there could be dangerous effects from AI improving bioweapons*, and because they're a relatively close analogue to AI by virtue of their dual-use, concealability, and reasonably large-scale effects.
Do you know of good sources on bioweapons strategy, offense-defense dynamics, and potential effects of future advances? I'm reading Koblentz's Living Weapons right now and it's quite good...
However, this approach is a bit silly because it does not model the acceleration of research: If there are no other donors in the field, then our donation is futile because £10,000 will not fund the entire effort required.
...Could you explain this more clearly to me please? With some stats as an example it'll likely be much clearer. Looking at the development of the Impossible Burger seems a fair phenomena to base GFI's model on, at least for now and at least insofar as it is being used to model a GFI donation's counterfactual impact in supporting simil
This is interesting. I'm strongly in favor of having rough models like this in general. Thanks for sharing!
Edit suggestions:
STI says "what percent of bad scenarios should we expect this to avert", but the formula uses it as a fraction. Probably best to keep the formula and change the wording.
Would help to clarify that TXR is a probability of X-risk. (This is clear after a little thought/inspection, but might as well make it as easy to use as possible.)
Quick thoughts:
It might be helpful to talk in terms of research-years rather than resea
Thanks for sharing! This seems like good news, and I'm glad they're looking at safety issues along so many different axes.
However, I'm a bit confused as to what interventions like this are meant/expected to accomplish. It seems like the long-term result of this kind of intervention would be a recovery of the mosquito population as the modified mosqs' descendents got outcompeted by mosquitos without the genes.
Is the idea that mosquito populations are small enough (relative to the number of modified ones introduced) that they might be eradicated entirely, ...
Hey! Just happened upon this article while searching for something else. Hope the necro isn't minded.
I wanted to point out that since this article was written--and especially in the last year--basic income at least has become a lot more mainstream. There's the (failed) Swiss referendum, and apparently Finland and YCombinator are both running basic income trials as well. (More locally, there's of course the GiveDirectly UBI trial as well.)
Anecdotally, it seems like these events have also been accompanied by many more people (in my particular bubble) being...
I've respected Rethink's work for a long time. Excited to see you guys expanding into the longtermist space!
Could you clarify how your stated "room for more funding" relates to your budget plans? For example, the maximally ambitious RFMF for longtermism in 2022 is $5.34m, but the maximally ambitious budget for 2022 is $2.5m. Is the idea that you would hold onto some money for 2023?