All of Thomas Kwa🔹's Comments + Replies

On a global scale I agree. My point is more that due to the salary standards in the industry, Eliezer isn't necessarily out of line in drawing $600k, and it's probably not much more than he could earn elsewhere; therefore the financial incentive is fairly weak compared to that of Mechanize or other AI capabilities companies.

3
NickLaing
It's true Mechanize are trying to hire him for 650k...
7
Ben Stevenson
Thanks for the reply. I agree with your specific point but I think it’s worth being more careful with your phrasing. How much we earn is an ethically-charged thing, and it’s not a good thing if EA’s relationship with AI companies gives us a permission structure to lose sight of this. Edit: to be clear, I agree that “it’s probably not  much more than he could earn elsewhere” but disagree that “Eliezer isn’t necessarily out of line in drawing $600k”

Being really good at your job is a good way to achieve impact in general, because your "impact above replacement" is what counts. If a replacement level employee who is barely worth hiring has productivity 100, and the average productivity is 150, the average employee will get 50 impact above replacement. If you do your job 1.67x better than average (250 productivity), you earn 150 impact above replacement, which is triple the average.

1
Vasco Grilo🔸
Thanks for the nice point, Thomas. Generalising, if the impact is 0 for productivity P_0, and P_av is the productivity of random employees, an employee N times as productive as random employees would be (N*P_av - P_0)/(P_av - P_0) as impactful as random employees. Assuming the cost of employing someone is proportional to their productivity, the cost-effectiveness as a fraction of that of random employees would be (P_av - P_0/N)/(P_av - P_0). So the cost-effectiveness of an infinitely productive employee as a fraction of that of random employees would be P_av/(P_av - P_0) = 1/(1 - P_0/P_av). In this model, super productive employees becoming more proctive would not increase their cost-effectiveness. It would just make them more impactful. For your parameters, employing an infinitely productive employee would be 3 (= 1/(1 - 100/150)) times as cost-effective as employing random employees.

I strongly disagree with a couple of claims:

MIRI's business model relies on the opposite narrative. MIRI pays Eliezer Yudkowsky $600,000 a year. It pays Nate Soares $235,000 a year. If they suddenly said that the risk of human extinction from AGI or superintelligence is extremely low, in all likelihood that money would dry up and Yudkowsky and Soares would be out of a job.

[...] The kind of work MIRI is doing and the kind of experience Yudkowsky and Soares have isn't really transferable to anything else.

  • $235K is not very much money [edit: in the context of
... (read more)
3
Yarrow Bouchard 🔸
If Mechanize succeeds in its long-term goal of "the automation of all valuable work in the economy", then everyone on Earth will be a billionaire.

$235K is not very much money. […] $600K is also not much money.


This is false.

$235K is not very much money. I made close to Nate's salary as basically an unproductive intern at MIRI.

I understand the point being made (Nate plausibly could get a pay rise from an accelerationist AI company in Silicon Valley, even if the work involved was pure safetywashing, because those companies have even deeper pockets), but I would stress that these two sentences underline just how lucrative peddling doom has become for MIRI[1] as well as how uniquely positioned all sides of the AI safety movement are.

There are not many organizations whose mes... (read more)

Is there a formula for the pledge somewhere? I couldn't find one.

5
Dan_Keys
The appendix of the ebook (pdf, p. 218 & 221) suggests structuring it like tax brackets, giving (as a percentage of adjusted gross income):
1
ethai
yeah not sure unfortunately!

See the gpt-5 report. "Working lower bound" is maybe too strong; maybe it's more accurate to describe it as an initial guess at a warning threshold for rogue replication and 10x uplift (if we can even measure time horizons that long). I don't know what the exact reasoning behind 40 hours was, but one fact is that humans can't really start viable companies using plans that only take a ~week of work. IMO if AIs could do the equivalent with only a 40 human hour time horizon and continuously evade detection, they'd need to use their own advantages and have made up many current disadvantages relative to humans (like being bad at adversarial and multi-agent settings).

What scale is the METR benchmark on? I see a line that "Scores are normalized such that 100% represents a 50% success rate on tasks requiring 8 human-expert hours.", but is the 0% point on the scale 0 hours?

METR does not think that 8 human hours is sufficient autonomy for takeover; in fact 40 hours is our working lower bound.

4
HugoSave
Indeed the 0%point is zero hours, so compared to the METR plot it is divided by 8 hours.  The 8 hours I agree is somewhat arbitrary and I had missed that METR had a more 'official' stance on it. I made an issue out of it now to see if anyone else had reasons to make it 8 hours. (For context I did most of the benchmark literature review for this project and data collection.)   Edit (29 Jan 2026): The change to 40 hour normalization is now live!
2
Denkenberger🔸
Could you please explain your reasoning on 40 hours?
8
David Mathers🔸
METR has an official internal view on what time horizons correspond to "takeover not ruled out"? 

What if we decide that the Amazon rainforest has a negative WAW sign? Would you be in favor of completely replacing it with a parking lot, if doing so could be done without undue suffering of the animals that already exist there?

Definitely not completely replacing because biodiversity has diminishing returns to land. If we pave the whole Amazon we'll probably extinct entire families (not to mention we probably cause ecological crises elsewhere and disrupt ecosystem services etc), whereas on the margin we'll only extinct species endemic to the deforested re... (read more)

It's plausible to me that biodiversity is valuable, but with AGI on the horizon it seems a lot cheaper in expectation to do more out-there interventions, like influencing AI companies to care about biodiversity (alongside wild animal welfare), recording the DNA of undiscovered rainforest species about to go extinct, and buying the cheapest land possible (middle of Siberia or Australian desert, not productive farmland). Then when the technology is available in a few decades and we're better at constructing stable ecosystems de novo, we can terraform the des... (read more)

I totally agree that there are some "out there" interventions that, in a perfect world, we would be funding much more. In particular biobanking (recording the DNA of species about to go extinct) should be considered much more, I totally agree. Unfortunately, the world is full of  techno-pessimists, deontologists, post-structuralists, diplomats who don't know what any of the preceding words even mean, etc. This seems insane, but MANY conservationists are against de-extinction for (in my view) fairly straightforward technophobic reasons. Convincing THOS... (read more)

2
David Mathers🔸
I don't think this is sufficient to explain EA disinterest, because there are also neartermist EAs who are skeptical about near-term AGI, or just don't incorporate it into their assessment of cause areas and interventions. 

A couple more "out-there" ideas for ecological interventions:

  • "recording the DNA of undiscovered rainforest species" -- yup, but it probably takes more than just DNA sequences on a USB drive to de-extinct a creature in the future.  For instance, probably you need to know about all kinds of epigenetic factors active in the embryo of the creature you're trying to revive.  To preserve this epigenetic info, it might be easiest to simply freeze physical tissue samples (especially gametes and/or embryos) instead of doing DNA sequencing.  You might
... (read more)

Thanks for the reply.

  • Everyone has different emotional reactions so I would be wary about generalizing here. Of the vegetarians I know, certainly not all are disgusted by meat. Disgust is often more correlated with whether they use a purity frame of morality or experience disgust in general than how much they empathize with animals [1]. Empathy is not an end; it's not required for virtuous action, and many people have utilitarian, justice-centered, or other frames that can prescribe actions with empathy taking a lesser role. As for me, I feel that after exp
... (read more)
1
Tristan Katz
To be clear, my reason for disgust isn't because I think that eating meat is impure. It's because seeing dead animals, or parts of dead animals, reminds me of the life that once existed and is now gone. This is the same reason for why I would never be appetized by human flesh - seeing dead people fills me with sorrow, and seeing many dead people in places where dead people shouldn't be (such as on supermarket shelves) fills me with horror, because it reminds me of the atrocity that continues. Many vegetarians don't have these emotions because they haven't fully recognized the atrocity for what it is. They aren't really coming to terms with the scale of suffering. Many vegans don't either, but I don't think it's possible to really be cognizant of the atrocity for what it is and not have this emotional reaction, for most people. For this reason, I want to be horrified, disgusted by meat, because I don't want to ignore the scale of the suffering around me. I want to be aware of and motivated by this wrong (except of course it is overwhelming and counterproductive). And I'm referring to myself here, but I think we would have much more pro-animal action if other people also tried to internalize nonspeciesm in this way too.  I hope that makes sense.

Didn't realize my only post of the year was from April 1st. Longforms are just so scary to write other than on April Fool's Day!

Are you interested in betting on these beliefs? I couldn't find a bet with Vasco but it seems more likely we could find one, because it seems like you're more confident

2
Yarrow Bouchard 🔸
It seems a bit redundant to make a bet on the stock market, no? But, for the heck of it: I’ll bet you a $10 donation to the charity of your/my choice that by January 1, 2031, at least four of the five following reputable news outlets declares the AI bubble has popped (or equivalent language that means the same thing): * The Wall Street Journal * The Financial Times * The Economist * Bloomberg * The New York Times If there’s any doubt or disagreement about whether the resolution criteria have been met when the time comes, I’d be willing to put it to an EA Forum poll or appoint a judge. If you want to counter-offer with a bet of your own devising, I’m all ears.
  • You're shooting the messenger. I'm not advocating for downvoting posts that smell of "the outgroup", just saying that this happens in most communities that are centered around an ideological or even methodological framework. It's a way you can be downvoted while still being correct, especially from the LEAST thoughtful 25% of EA forum voters
  • Please read the quote from Claude more carefully. MacAskill is not an "anti-utilitarian" who thinks consequentialism is "fundamentally misguided", he's the moral uncertainty guy. The moral parliament usually recommends actions similar to consequentialism with side constraints in practice.

I probably won't engage more with this conversation.

Claude thinks possible outgroups include the following, which is similar to what I had in mind

Based on the EA Forum's general orientation, here are five individuals/groups whose characteristic opinions would likely face downvotes:

  1. Effective accelerationists (e/acc) - Advocates for rapid AI development with minimal safety precautions, viewing existential risk concerns as overblown or counterproductive
  2. TESCREAL critics (like Emile Torres, as you mentioned) - Scholars who frame longtermism/EA as ideologically dangerous, often linking it to eugenics, colonialism
... (read more)
2
Yarrow Bouchard 🔸
a) I’m not sure all of those count as someone who would necessarily be an outsider to EA (e.g. Will MacAskill only assigns a 50% probability to consequentialism being correct, and he and others in EA have long emphasized pluralism about normative ethical theories; there’s been an EA system change group on Facebook since 2015 and discourse around systemic change has been happening in EA since before then) b) Even if you do consider people in all those categories to be outsiders to EA or part of "the out-group", us/them or in-group/out-group thinking seems like a bad idea, possibly leading to insularity, incuriosity, and overconfidence in wrong views  c) It’s especially a bad idea to not only think in in-group/out-group terms and seek to shut down perspectives of "the out-group" but also to cast suspicion on the in-group/out-group status of anyone in an EA context who you happen to disagree with about something, even something minor — that seems like a morally, subculturally, and epistemically bankrupt approach 
  • My main concern is that the arrival of AGI completely changes the situation in some unexpected way.
    • e.g. in the recent 80k podcast on fertility, Rob Wiblin opines that the fertility crash would be a global priority if not for AI likely replacing human labor soon and obviating the need for countries to have large human populations. There could be other effects.
    • My guess is that due to advanced AI, both artificial wombs and immortality will be technically feasible in the next 40 years, as well as other crazy healthcare tech. This is not an uncommon view
  • Before
... (read more)
3
David Mathers🔸
"Rob Wiblin opines that the fertility crash would be a global priority if not for AI likely replacing human labor soon and obviating the need for countries to have large human populations" This is a case where it really matters whether you are giving an extremely high chance that AGI is coming within 20-30 years, or merely a decently high chance. If you think the chance is like 75%, and the claim that conditional on no AGI, low fertility would be a big problem is correct, then the problem is only cut by 4x, which is compatible with it still being large and worth working on. Really, you need to get above 97-8% before it starts looking clear that low fertility is not worth worrying about, if we assume that conditional on no AGI it will be a big problem. 

Yeah, while I think truth-seeking is a real thing I agree it's often hard to judge in practice and vulnerable to being a weasel word.

Basically I have two concerns with deferring to experts. First is that when the world lacks people with true subject matter expertise, whoever has the most prestige--maybe not CEOs but certainly mainstream researchers on slightly related questions-- will be seen as experts and we will need to worry about deferring to them.

Second, because EA topics are selected for being too weird/unpopular to attract mainstream attention/fund... (read more)

Not "everyone agrees" what "utilitarianism" means either and it remains a useful word. In context you can tell I mean someone whose attitude, methods and incentives allow them to avoid the biases I listed and others.

1
Yarrow Bouchard 🔸
If I want to know what “utilitarianism” means, including any disagreements among scholars about the meaning of the term (I have a philosophy degree, I have studied ethics, and I don’t have the impression there are meaningful disagreements among philosophers on the definition of “utilitarianism”), I can find this information in many places, such as: * The Stanford Encyclopedia of Philosophy * Encyclopedia Britannica * Wikipedia * The book Utilitarianism: A Very Short Introduction co-authored by Peter Singer and published by Oxford University Press * A textbook like Normative Ethics or an anthology like Ethical Theory * Philosophy journals * An academic philosophy podcast like Philosophy Bites * Academic lectures on YouTube and Crash Course (a high-quality educational resource) So, it’s easy for me to find out what “utilitarianism” means. There is no shortage of information about that. Where do I go to find out what “truth-seeking” means? Even if some people disagree on the definition, can I go somewhere and read about, say, the top 3 most popular definitions of the term and why people prefer one definition over the other?  It seems like an important word. I notice people keep using it. So, what does it mean? Where has it been defined? Is there a source you can cite that attempts to define it? I have tried to find a definition for “truth-seeking” before, more than once. I’ve asked what the definition is before, more than once. I don’t know if there is a definition. I don’t know if the term means anything definite and specific. I imagine it probably doesn’t have a clear definition or meaning, and that different people who say “truth-seeking” mean different things when they say it — and so people are largely talking past each other when they use this term.  Incidentally, I think what I just said about “truth-seeking” probably also largely applies to “epistemics”. I suspect “epistemics” probably either means epistemic practices or epistemology, but it’s not

I think the "most topics" thing is ambiguous. There are some topics on which mainstream experts tend to be correct and some on which they're wrong, and although expertise is valuable on topics experts think about, they might be wrong on most topics central to EA. [1] Do we really wish we deferred to the CEO of PETA on what animal welfare interventions are best? EAs built that field in the last 15 years far beyond what "experts" knew before.

In the real world, assuming we have more than five minutes to think about a question, we shouldn't "defer" to experts ... (read more)

4
Arepo
I agree with Yarrow's anti-'truth-seeking' sentiment here. That phrase seems to primarily serve as an epistemic deflection device indicating 'someone whose views I don't want to take seriously and don't want to justify not taking seriously'. I agree we shouldn't defer to the CEO of PETA, but CEOs aren't - often by their own admission - subject matter experts so much as people who can move stuff forwards. In my book the set of actual experts is certainly murky, but includes academics, researchers, sometimes forecasters, sometimes technical workers - sometimes CEOs but only in particular cases - anyone who's spent several years researching the subject in question.  Sometimes, as you say, they don't exist, but in such cases we don't need to worry about deferring to them. When they do, it seems foolish to not to upweight their views relative to our own unless we've done the same, or unless we have very concrete reasons to think they're inept or systemically biased (and perhaps even then).
-9
Yarrow Bouchard 🔸

I think this is a significant reason why people downvote some, but not all, things they disagree with. Especially a member of the outgroup who makes arguments EAs have refuted before and need to reexplain, not saying it's actually you

2
Yarrow Bouchard 🔸
What is "the outgroup"?

Can you explain what you mean by "contextualizing more"? (What a curiously recursive question...)

I mean it in this sense; making people think you're not part of the outgroup and don't have objectionable beliefs related to the ones you actually hold, in whatever way is sensible and honest.

Maybe LW is better at using disagreement button as I find it's pretty common for unpopular opinions to get lots of upvotes and disagree votes. One could use the API to see if the correlations are different there.

2
Yarrow Bouchard 🔸
Huh? Why would it matter whether or not I'm part of "the outgroup"...? What does that mean?

IMO the real answer is that veganism is not an essential part of EA philosophy, just happens to be correlated with it due to the large number of people in animal advocacy. Most EA vegans and non-vegans think that their diet is a small portion of their impact compared to their career, and it's not even close! Every time you spend an extra $5 finding a restaurant with a vegan option you could help 5,000 shrimp instead. Vegans have other reasons like non-consequentialist ethics, virtue signaling or self-signaling, or just a desire not to eat the actual flesh/... (read more)

I really enjoyed reading this post; thanks for writing it. I think it's important to take space colonization seriously and shift into "near mode" given that, as you say, the first entity to start a Dyson Swarm has a high chance to get DSA if it isn't already decided by AGI, and it's probably only 10-35 years away.

Regarding COIs, it's probably bigger that Daniela is married to Holden, and while not strictly a COI, we don't want the association with OP's political advocacy. There are probably other things, I'm don't work on strategy

Assorted thoughts

  • Rate limits should not apply to comments on your own quick takes
  • Rate limits could maybe not count negative karma below -10 or so, it seems much better to rate limit someone only when they have multiple downvoted comments  
  • 2.4:1 is not a very high karma:submission ratio. I have 10:1 even if you exclude the april fool's day posts, though that could be because I have more popular opinions, which means that I could double my comment rate and get -1 karma on the extras and still be at 3.5
  • if I were Yarrow I would contextualize more or use m
... (read more)
2
Yarrow Bouchard 🔸
Can you explain what you mean by "contextualizing more"? (What a curiously recursive question...) You definitely have more popular opinions (among the EA Forum audience), and also you seem to court controversy less, i.e. a lot of your posts are about topics that aren't controversial on the EA Forum. For example, if you were to make a pseudonymous account and write posts/comments arguing that near-term AGI is highly unlikely, I think you would definitely get a much lower karma to submission ratio, even if you put just as much effort and care into them as the posts/comments you've written on the forum so far. Do you think it wouldn't turn out that way? I've been downvoted on things that are clearly correct, e.g. the standard definitions of terms in machine learning (which anyone can Google); a methodological error that the Forecasting Research Institute later acknowledged was correct and revised their research to reflect. In other cases, the claims are controversial, but they are also claims where prominent AI experts like Andrej Karpathy, Yann LeCun, or Ilya Sutskever have said exactly the same thing as I said — and, indeed, in some cases I'm literally citing them — and it would be wild to think these sort of claims are below the quality threshold for the EA Forum. I think that should make you question whether downvotes are a reliable guide to the quality of contributions. One-off instances of one person downvoting don't bother me that much — that literally doesn't matter, as long as it really is one-off — what bothers me is the pattern. It isn't just with my posts/comments, either, it's across the board on the forum. I see it all the time with other contributors as well. I feel uneasy dragging those people into this discussion without their permission — it's easier to talk about myself — but this is an overall pattern. Whether reasoning is good or bad is always bound to be controversial when debating about topics that are controversial, about which there is a lo

My understanding is METR doesn't take Good Ventures money to avoid the appearance of COIs. We could maybe avoid creating actual COIs but it is crucial to the business model to appear as trusted and neutral as possible.

2
Yarrow Bouchard 🔸
What's the risk of an appearance that there's a COI? Is it that Dustin Moskovitz is an Anthropic investor?

When 80,000 Hours pivoted to AI, I largely stopped listening to the podcast, thinking that as part of the industry I would already know everything. But I recently found myself driving a lot and consuming more audio content, and the recent ones eg with Holden, Daniel K and ASB are incredibly high quality and contain highly nontrivial, grounded opinions. If they keep this up I will probably keep listening until the end times.

What inspiring and practical examples!

Maybe a commitment to impact causes EA parents to cooperate at maximizing it, which means optimally distributing the parenting workload whatever society thinks. In EA with lots of conferences and hardworking impactful women, it makes sense that the man's op cost is often lower. Elsewhere couples cooperate to maximize income, but men tend to have higher earning potential so maybe the woman would often do more childcare anyway.

My sense is that parenting falls on the woman due not only to gender norms, but also higher ave... (read more)

I don't have reason to think that prioritizing women's careers is more common in EA than in other similarly educated groups. And within EA, I definitely think it's still most common that women are doing more of the parenting work. But I wanted to highlight some examples to show that multiple configurations really are possible!

There are a few mistakes/gaps in the quantitative claims:

Continuity: If A ≻ B ≻ C, there's some probability p ∈ (0, 1) where a guaranteed state of the world B is ex ante morally equivalent to "lottery p·A + (1-p)·C” (i.e., p chance of state of the world A, and the rest of the probability mass of C)

This is not quite the same as either property 3 or property 3' in the Wikipedia article, and it's plausible but unclear to me that you can prove 3' from it. Property 3 uses "p ∈ [0, 1]" and 3' has an inequality; it seems like the argument still goes through with ... (read more)

1
Aaron Bergman
Thanks, yeah I may have gotten slightly confused when writing. 1) VNM Wikipedia screenshot: Let P be the thing I said in the post: or, symbolically P≡(A≻B≻C⇒∃p∈(0,1)[B∼pA+(1−p)C]) and let Q≡(L≺M≺N⇒∃ε∈(0,1)[(1−ε)L+εN≺M≺εL+(1−ε)N]) I think (P and Independence)⟹Q but not P⟹Q in general. So my writing was sloppy. Super good catch (not caught by any of the various LLMs iirc!) But for the purposes of the argument everything holds together because you need independence axiom for VNM to hold. But still, sloppy.  2) "grows without bound" bit Me: "This arbitrariness diminishes somewhat (though, again, not entirely) when viewed through the asymptotic structure. Once we accept that compensation requirements grow without bound as suffering intensifies, some threshold becomes inevitable. The asymptote must diverge somewhere; debates about exactly where are secondary to recognizing the underlying pattern." You: Straightforward error by me, I will change the wording. Not sure how that happened 3) "continuity" Yeah, idk, English and math only provide so many words. I could have spent more words more driving home and clarifying this point or inventing and defining additional terms. My intuition is that it's clear enough as is (evidently we disagree about this) but if a couple other people say "yeah this is misleading and confusing" then I'll concede that I made a bad choice about clarity vs brevity as a writing decision. 4) "write down", "compute", and "physically instantiate" till end Ngl I am pretty confused about everything starting here. I think I'm just reading you wrong somehow. Like the difference in those magnitudes is huge, point taken, but I don't see why that matters for my argument.  Confused here because yeah clearly adding t+epsilon and removing t-epsilon gives you a net change below zero. But I sense you might be getting at the (very substantive and important) cluster of critiques I respond to in this comment (?) Yeah I'm ~totally agnostic about

Ok interesting! I'd be interested in seeing this mapped out a bit more, because it does sound weird to have BOS be offsettable with positive wellbeing, positive wellbeing to be not offsettable with NOS, but BOS and NOS are offsetable with each other? Or maybe this isn't your claim and I'm misunderstanding 

This is what kills the proposal IMO, and EJT also pointed this out. The key difference between this proposal and standard utilitarianism where anything is offsettable isn't the claim that that NOS is worse than TREE(3) or even 10^100 happy lives, sin... (read more)

It is not necessary to be permanently vegan for this. I have only avoided chicken for about 4 years, and hit all of these benefits.

  • Because evidence suggests that when we eat animals we are likely to view them as having lower cognitive capabilities or moral status (see here for a wikipedia blurb about it).
    • I have felt sufficient empathy for chicken for basically the whole time I haven't eaten it. I also went vegan for (secular) Lent four years ago, and felt somewhat more empathy for other animals, but my sense is eating non-chicken animals didn't meaningfull
... (read more)
3
Tristan Katz
Sorry I missed this comment when you originally made it. These are good arguments, but I disagree that you're hitting the same benefits. They're the same kind of reasons for adopting that diet, but the benefits are different: * Even if you consider your empathy 'sufficient', it's not clear you're feeling the same empathy. Speaking anecdotally: I went vegetarian and vegan gradually, bit by bit, and I know that something clicked in my mind when I gave up meat altogether. If you really cared about other animals and saw them as conscious beings, you should be disgusted by meat, in a similar way to how you would feel about eating human meat. You should be horrified - at least, if your emotions work the way they do for most humans. So assuming that you're not highly neurodivergent in a way that makes you the exception to this rule, I assume that you're not empathising with animals at the same level. This has implications for how you do cause prioritisation, where you donate, what career you choose etc. * You're giving a different social signal (that eating chicken is wrong, rather than eating meat is wrong). * I'll agree with the third point, not eating chicken is clearly easier. Regarding the last point, I would agree... except that supply of meat is increasing, both globally and in the west. Efforts to change the minds of individual people aren't working, and won't work as long as structural issues e.g. the subsidisation of meat continue. 
3
Suzanne
All I want to add is that pigs suffer some of the most suffering possible in my opinion. If you are trying to limit consumption of the worst foods it might be worth looking more into the standard practices they endure (gestation crates, farrowing crates, and teeth clipping and tail docking without anesthetic).

I perceive it as +EV to me but I feel like I'm not the best buyer of short timelines. I would maybe do even odds on before 2045 for smaller amounts, which is still good for you if you think the yearly chance won't increase much. Otherwise maybe you should seek a bet with someone like Eli Lifland. The reason I'm not inclined to make large bets is that the markets would probably give better odds for something that unlikely, eg options that pay out with very high real interest rates; whereas a few hundred dollars is enough to generate good EA forum discussion.

2
Vasco Grilo🔸
Thanks, Thomas. @elifland, would you like to make the bet I mentioned above (you should expect to gain money if you think your probability of winning is higher than 25 %)?

No bet. I don't have a strong view on short timelines or unemployment. We may find a bet about something else; here are some beliefs

  • my or Linch's position vs yours on probability of extinction from nuclear war (I'd put $2 against your $98 that you ever update upwards by 50:1 on extinction by nuclear war by 2050, but no more for obvious reasons)
  • >25% that global energy consumption will increase by 25% year over year some year before 2035 (30% is the AI expert median, superforecaster median is <1%), maybe more
  • probably >50% that a benchmark by Mechani
... (read more)

Thanks for the suggestions, Thomas! Would you like to make the following bet?

  • If the global energy consumption as reported by Our World in Data (OWID) increases by more than 25 % from year Y to Y + 1, for any Y <= 2033 (such that the maximum Y + 1 is 2034, before 2035), I donate 12 k$ to an organisation or fund of your choice in July 2035 (OWID updates their data in June).
  • Otherwise, you donate 4 k$ to an organisation or fund of my choice in July 2035.

Your expected gain is more than 0 (= (0.25*12 - (1 - 0.25)*4)*10^3), but sufficiently so? You said ">2... (read more)

Agree. Given that Vasco is willing to give 2:1 odds for 2029 below, this bet should have been 3:1 or better for David. It would have been a better signal of the midpoint odds to the community.

4
Davidmanheim
I would have been happy to get better odds, obviously! 

A footnote says the 0.15% number isn't an actual forecast: "Participants were asked to indicate their intuitive impression of this risk, rather than develop a detailed forecast". But superforecasters' other forecasts are roughly consistent with 0.15% for extinction, so it still bears explaining.

In general I think superforecasters tend to anchor on historical trends, while AI safety people anchor on what's physically possible or conceivable. Superforecasters get good accuracy compared to domain experts on most questions because domain experts in many fields... (read more)

1
PSR
Thank you a lot for this detailed answer. Especially points where superforecasters have provably been wrong on AI-related questions are very interesting and are certainly a very relevant argument against updating too much in their direction. Some kind of track record of superforecasters, experts, and public figures making predictions would be extremely interesting. Do you know whether something like this can be found somewhere? To push back a bit against it being hard to find a good reference class and superforecasters having to rely on vibes: Yes, it might be hard, but aren't superforecasters precisely those who have a great track record for finding a good methodology for making predictions, even when it's hard? AI extinction is probably not the only question where making a forecast is tricky. Edit: Just a few days ago, we got this here, which is very relevant: https://forum.effectivealtruism.org/posts/fp5kEpBkhWsGgWu2D/assessing-near-term-accuracy-in-the-existential-risk

I think I would take your side here. Unemployment above 8% requires replacing so many jobs that humans can't find new ones elsewhere even during the economic upswing created by AGI, and there is less than 2 years until the middle of 2027. This is not enough time for robotics (on current trends robotics time horizons will be under 1 hour) and AI companies can afford to keep hiring humans even if they wouldn't generate enough value most places, so the question is whether we see extremely high unemployment in remotable sectors that automate away existing jobs... (read more)

4
Vasco Grilo🔸
Thanks for sharing your thoughts, Thomas! Would you like to make the following bet? * If the mean monthly unemployment rate in the US in 2029, as reported by FRED, is higher than 8 %, I donate 10 k$ to an organisation or fund of your choice in January 2030. * Otherwise, you donate 5 k$ to an organisation or fund of my choice in January 2030.
9
Davidmanheim
It's partly shorter timelines, which we're seeing start to play out, and partly underlying pessimism on US economic policy under Trump, and the increasing odds of a recession.  The Us economy has been stalled, and the only reason this isn't obvious in the stock market is the AI companies - so my weak general model is that either the AI companies continue to do better, which at least weakly implies job displacement, or they don't, and there's a market crash and need for stimulus and inflation. In that situation, or even with a continued status quo maybe AI matters/maybe it doesn't, then absent a pretty strong recovery elsewhere - which seems unlikely given the ongoing uncertainty under Trump - implies that we could have a pretty major setback in the broader economy.

Spreading around the term “humane meat” may get it into some people’s heads that this practice can be humane, which could in turn increase consumption overall, and effectively cancel out whatever benefits you’re speculating about. 

I don't know what the correct definition of "humane" is, but I strongly disagree with this claim in the second half. The question is whether higher-welfare imports reduce total suffering once we account for demand effects. So we should care about improving conditions from "torture camps" -> "prisons" -> "decent". Tortu... (read more)

Yeah, because I believe in EA and not in the socialist revolution, I must believe that EA could win some objective contest of ideas over socialism. In the particular contest of EA -> socialist vs socialist -> EA conversions I do think EA would win since it's had a higher growth rate in the period both existed, though someone would have to check how many EA deconverts from the FTX scandal became socialists. This would be from both signal and noise factors; here's my wild guess at the most important factors:

  • Convinced EAs have been exposed to the "root
... (read more)

At risk of further psychoanalyzing the author, it seems like they're naturally more convinced by forms of evidence that EAs use, and had just not encountered them until this project. Many people find different arguments more compelling, either because they genuinely have moral or empirical assumptions incompatible with EA, or because they're innumerate. So I don't think EA has won some kind of objective contest of ideas here.

Nevertheless this was an interesting read and the author seems very thoughtful.

4
Linch
This was my thoughts on it as well, along with some more general pondering of the extent to which methodological bias (on the level of selecting methods with built-in directional bias) affects our ability to understand the truth, even when we try very hard.
9
Karthik Tadepalli
Yes, nothing in this post seems less likely than an EA trying to convince socialists to become EAs and subsequently being convinced of socialism.

One problem is putting everything on a common scale when historical improvements are so sensitive to the distribution of tasks. A human with a computer with C, compared to a human with just log tables, is a billion times faster at multiplying numbers but less than twice as fast at writing a novel. So your distribution of tasks has to be broad enough that it captures the capabilities you care about, but it also must be possible to measure a baseline score at low tech level and have a wide range of possible scores. This would make the benchmark extremely difficult to construct in practice.

2
Arepo
I think that's right, but modern AI benchmarks seem to have much the same issue. A human with a modern Claude instance might be able to write code 100x faster than without, but probably less than 2x as fast at choosing a birthday present for a friend. Ideally you want to integrate over... something to do with the set of all tasks. But it's hard to say what that something would be, let alone how you're going to meaningfully integrate it.

If your algorithms get more efficient over time at both small and large scales, and experiments test incremental improvements to architecture or data, then they should get cheaper to run proportionally to algorithmic efficiency of cognitive labor. I think this is better as a first approximation than assuming they're constant, and might hold in practice especially when you can target small-scale algorithmic improvements.

4
Owen Cotton-Barratt
OK I see the model there. I guess it's not clear to me if that should hold if I think that most experiment compute will be ~training, and most cognitive labour compute will be ~inference? However, over time maybe more experiment compute will be ~inference, as it shifts more to being about producing data rather than testing architectures? That could push back towards this being a reasonable assumption. (Definitely don't feel like I have a clear picture of the dynamics here, though.)

I'm worried that trying to estimate  by looking at wages is subject to lots of noise due to assumptions being violated, which could result in the large discrepancy you see between the two estimates.

One worry: I would guess that Anthropic could derive more output from extra researchers (1.5x/doubling?) than from extra GPUs (1.18x/doubling?), yet it spends more on compute than researchers. In particular I'd guess alpha/beta = 2.5, and wages/r_{research} is around 0.28 (maybe you have better data here). Under Cobb-Douglas and perfect competition th... (read more)

7
Parker_Whitfill
Thanks for the insightful comment.  I take your overall point as the static optimization problem may not be properly specified. For example, costs may not be linear in labor size because of adjustment costs to growing very quickly or costs may not be linear in compute because of bulk discounting. Moreover, these non-linear costs may be changing over time (e.g., adjustment costs might only matter in 2021-2024 as OpenAI, Anthropic have been scaling labor aggressively). I agree that this would bias the estimate of σ. Given the data we have, there should be some way to at least partially deal with this (e.g., by adding lagged labor as a control). I'll have to think about it more.  On some of the smaller comments:  The best data we have is The Information's article that OpenAI spent $700M on salaries and $1000M on research compute in 2024, so the wLrK=.7 (assuming you meant wLrK instead of wr).  I agree. σ might not be constant over time, which is a problem for both estimation/extrapolation and also predicting what an intelligence explosion might look like. For example, if σ falls over time, then we may have a foom for a bit until σ falls below 1 and then fizzles. I've been thinking about writing something up about this.  Yes, although we're not decided yet on what is the most useful to follow-up on. Very short-term there is trying to accomodate non-linear pricing. Of course, data on what non-linear pricing looks like would be helpful e.g., how does Nvidia bulk discount.  We also may try to estimate ϕ with the data we have. 

(edited to fix numbers, I forgot 2 boxes means +3dB)

dB is logarithmic so a proportional reduction in sound energy will mean subtracting an absolute number of dB, not a percentage reduction in dB.

HouseFresh tested the AirFanta 3Pro https://housefresh.com/airfanta-3pro-review/ at different voltage levels and found:

  • 12.6 V: 56.3 dBA, 14 minutes
  • 6.54 V: 43.3 dBA, 28 minutes

So basically you subtract 13 dB when halving the CADR. I now realize that if you have two boxes, the sound energy will double (+3dB) and so you'll actually only get -10 dB from running two at ... (read more)

1
JesseSmith
Bummer. Operating at 25 db would have been really impressive for a room filter but it looks like it's not even close. Extrapolating from David Elfstrom's airflow estimates it looks like ~1/2 cfm would put it >40 db, which isn't great. So if we want to create a satisfactory occupant experience we probably shouldn't put down the tin snips just yet.

FWIW I predict they will be a constant factor harder but improve at similar rates. Any particular benchmarks you think I should look at?

2
Vasco Grilo🔸
I do not have any particular benchmarks in mind, but I think the ones Mechanize is aiming to develop will capture economic value more closely. So you may want to look at the existing benchmarks which you think will more closely resemble those. I would also ask Mechanize about it, @Thomas Kwa. Your top comment no longer includes the graph.

I broadly agree with section 1, and in fact since we published I've been looking into how time horizon varies between domains. Not only is there lots of variance in time horizon, the rate of increase also varies significantly. 

See a preliminary graph plus further observations on LessWrong shortform.

2
Vasco Grilo🔸
Thanks for sharing, Thomas! I expect benchmarks whose scores are closer to being proportional to economic output improve slower.

You can get down to 25 dB by running two at half speed. Fan noise is proportional to RPM^5, so 50% speed will mean -15dB noise. The fans just need enough static pressure to maintain close to 50% airflow at 50% speed.

3
JesseSmith
Do you have tested numbers? Going from high to low on 4 filter CR boxes shows a reduction in airflow in the range of 30%-40% and db reduction around ~13%.

Usage varies-- the top five posts on /r/crboxes all use PC fans. Other guides do too,  and CleanAirKits and Nukit both describe themselves as PC fan CR boxes.

2
Jeff Kaufman 🔸
Interesting! I hadn't realized people had started using the term this way!

That's a box fan CR box; the better design (and the one linked) uses PC fans which are better optimized for noise. I don't have much first-hand experience with this, but physics suggests that noise from the fan will be proportional to power usage, which is pressure * airflow, if efficiency is constant, and this is roughly consistent with various tests I've found online.

Both further upsizing and better sound isolation would be great. What's the best way to reduce duct noise in practice? Is an 8" flexible duct quieter than a 6" rigid duct or will most of the noise improvement come from oversizing the termination, removing tight bends or installing some kind of silencer device? I might suggest this to a relative.

2
Jeff Kaufman 🔸
A filter with PC fans isn't a "CR box" -- a Corsi-Rosenthal box is specifically a design based on box fans. For example, Wikipedia, US Davis, Clean Air Crew, and the Corsi-Rosenthal Foundation mention only box fans.
3
JesseSmith
Ah, sorry, the CR reference threw me off. That link seems to suggest those are in the 40 db range.  Duct design uses equivalent length as a metric across duct types and fittings with 1 foot of straight metal pipe as a '1'. The problem with flex is that each foot has an equivalent length of 1.5, which adds a moderate amount to system length (although to keep this in perspective many near air handler transitions are >100'). For something like this I'd probably do a short run of flex near the motor (both sides) to dampen motor/vibration noise and transition to metal beyond. Avoid cornering with flex! In general, remote fans, eliminating line of sight on the motor and oversizing ducts/terminations will be shockingly quiet to most room occupants. If you want to get super clever, you can start doing offset openings in walls by cutting grilles high on one side and low on the other, but it can be difficult to pull sufficient air while keeping velocity low (400 cfm = 6 stud bays).

Isn't particulate what we care about? The purpose of the filters is to get particulate out of the air, and the controlled experiment Jeff did basically measures that. If air mixing is the concern, ceiling fans can mix air far more than required, and you can just measure particulate in several locations anyway.

3
JesseSmith
It's an OK experiment, but there could be other reasons for air exchange in the room that vary quite widely. Building infiltration would exchange air with outdoors and varies quite a bit with outdoor conditions. It's also not very difficult to measure airflow so I don't see many disadvantages to doing so

A pair of CR boxes can also get 350 CFM CADR at the same noise level for less materials cost than either this or the ceiling fan, and also have much less installation cost. E.g. two of this CleanAirKits model on half speed would probably cost <$250 if it were mass-produced. This is the setup in my group house living room and it works great! DIY CR boxes can get to $250/350 CFM right now.

The key is having enough filter area to make the static pressure and thus power and noise minimal-- the scaling works out such that every doubling of filter area at a gi... (read more)

3
JesseSmith
Are you limiting your noise estimates to strictly airflow/static pressure? The research I'm aware of suggests Corsi boxes perform at >40db and up to 60db on high speed (~350 cfm). https://housefresh.com/corsi-rosenthal-box-review . During the pandemic this was widely discussed as a major impediment to their adoption.  Upsizing terminations substantially reduces room-facing air velocity. Assuming 400 cfm, the velocity at the termination I link will be roughly 175 fpm. This is extremely low, and I'd expect to achieve 20-25 db. Most residential return registers are undersized and thus outside ACCA's duct design standard of 500 fpm. It's also less than standards widely used for quiet spaces (200-300 fpm). You could continue pursuing improvements here via upsizing further if you're so inclined, or using clever placement to put the filtration system away from the room activities or behind (offset) cabinetry or furnishings. 

I'm a big fan of this. Imagine if this becomes the primary way billionaires are ranked on prestige

Load more