All of Remmelt's Comments + Replies

Yeah, it's a case of people being manipulated into harmful actions. I'm saying 'besides' because it feels like a different category of social situation than seeing someone take some public action online and deciding for yourself to take action too.

1
Matrice Jacobine🔸🏳️‍⚧️
One of the killings was, as far as we know, purely mimetic and (allegedly) made by someone (@Maximilian Snyder) who never even interacted online with Ziz, so I don't think it's an invalid example to bring up actually.

My reaction here was: 'Good, someone shows they care enough about this issue that they're willing to give a costly signal to others that this needs to be taken seriously' (i.e. your point a).

I do personally think many people in EA and rationalist circles (particularly those concerned about AI risk) can act more proactively to try and prevent harmful AI developments (in non-violent ways). 

It's fair though to raise the concern that Guido's hungerstrike could set an example for others to take actions that are harmful to themselves. If you have any exampl... (read more)

1
Matrice Jacobine🔸🏳️‍⚧️
This is a very big "besides"!

I just expanded the text:

On one hand, it was a major contribution for a leading AI company to speak out against the moratorium as stipulated. On the other hand, Dario started advocating himself for minimal regulation. He recommended mandating a transparency standard along the lines of RSPs, adding that state laws "should also be narrowly focused on transparency and not overly prescriptive or burdensome".[11] Given that Anthropic had originally described SB 1047's requirements as 'prescriptive' and 'burdensome', Dario was effectively arguing for the fe

... (read more)

You’re right. I totally skipped over this.

Let me try to integrate that quote into this post. 

2
Remmelt
I just expanded the text:

FAAI evolution can happen horizontally and rapidly, unlike biological evolution

Note that horizontal code transfer can happen under biological evolution too. E.g. with bacteria. 

For the rest, this summary is roughly accurate!

I adjusted my guesstimate of winning down to a quarter.

I now guess it's more like 1/8 chance (meaning that from my perspective Marcus will win this bet on expectation). It is pretty hard to imagine so many paying customers going away, particularly as revenues have been growing in the last year.

Marcus has thought this one through carefully, and I'm naturally sticking to the commitment. If we end up seeing a crash down the line, I invite all of you to consider with me how to make maximum use of that opportunity!

I still think a crash is fairly likely, but als... (read more)

like AI & ML VC deal activity being <30% and Anthropic valuation <$30B 

My preference was for the former metric (based on AI PitchBook-NVCA Venture Monitor), and another metric based on some threshold for the absolute amount Anthropic or OpenAI got in investments in a next round (which Marcus reasonably pointed out could be triggered if the company just decided to do a some extra top-up round).

I was okay with using Marcus’ Anthropic valuation metric with the threshold set higher, and combined with another possible metric. My worry was that An... (read more)

Right, I don’t have a high income, and also have things in my personal life to take care of. 

Good question. 

Marcus and I did a lot of back and forth on potential criteria. I started by suggesting metrics that capture a decline in investments into AI companies. Marcus though was reasonably trying to avoid things that can be interest rate/tariff/broad market driven.

So the criteria we have here are a result of compromise.

The revenue criteria are rather indirect for capturing my view on things. I think if OpenAI and Anthropic each continue to make $5+ billion yearly losses (along with losses by other model developers) that would result in investo... (read more)

2
Remmelt
Here is the post about our bet.

It’s bad to support a race here. Given that no-one has a way to safely constrain open-endedly learning autonomous machinery, and there are actual limits to control. 

1
Tristan D
Yes I agree it's bad to support a race, but it's not as simple as that.

Apr: Californian civil society nonprofits

This petition has the most rigorous legal arguments in my opinion.
 

Others I know also back a block (#JusticeForSuchir, Ed Zitron, Stop AI, creatives for copyright). What’s cool is how diverse the backers are, from skeptics to doomers, and from tech whistleblowers to creatives. 

0
Tristan D
Is OpenAI going for profit in order to attract more investment/talent a good or bad thing for AI safety?  On the one hand people want American companies to win the AGI race, and this could contribute to that. On the other hand, OpenAI would be then be more tied to making profit which could conflict with AI safety goals.

Frankly, because I'd want to profit from it.

The odds of 1:7 imply a 12.5% chance of a crash, and I think the chance is much higher (elsewhere I posted a guess of 40% for this year, though I did not have precise crash criteria in mind there, and would lower the percentage once it's judged by a few measures, rather than my sense of "that looks like a crash"). 

That percentage of 12.5% is far outside of the consensus on this Metaculus page. Though I notice that their criteria for a "bust or winter" are much stricter than where I'd set the threshold for a ... (read more)

Haha, I was thinking about that. The timing was unfortunate. 

Just because they didn't invest money in the Stargate expansion doesn't mean they aren't reserving the option to do so later if necessary.... If you believe that the US government will prop up AI companies to virtually any level they might realistically need by 2029, then i don't see a crash happening.


Thanks for reading and your thoughts.

I disagree, but I want to be open to changing my mind if we see e.g. the US military ramping up contracts, or the US government propping up AI companies with funding at the level of say the $280 billion CHIPS Act.

This is clarifying context, thanks. It's a common strategy to go red for years while tech start-ups build a moat around themselves (particularly through network effects). Amazon built a moat in terms of drawing in vendors and buyers into its platform while reducing logistics costs, and Uber in drawing in taxi drivers and riders onto its platform. Tesla started out with a technological edge. 

Currently, I don't see a strong case for that OpenAI and Anthropic are building up a moat.
–> Do you have any moats in mind that I missed? Curious.

Network effect... (read more)

4
Ozzie Gooen
I think there are serious risks for LLM development (i.e. a better DeepSeek could be released at any points), but also some serious opportunities. 1. The game is still early. It's hard to say what moats might exist 5 years from now. This is a chaotic field. 2. ChatGPT/Claude spend a lot of attention on their frontends, the API support, documentation, monitoring, moderation, lots of surrounding tooling. It's a ton of work to make a high production-grade service, besides just having one narrow good LLM. 3. There's always the chance of something like a Decisive Strategic Advantage later.  Personally, if I were an investor, both would seem promising to me. Both are very risky - high chances of total failure, depending on how things play out. But that's common for startups. I'd bet that there's a good chance that moats will emerge later. 

Ah, it's meant to be the footnotes. Let me edit that to be less confusing.

Update: back up to 70% chance.

Just spent two hours compiling different contributing factors. Now I weighed those factors up more comprehensively, I don't expect to change my prediction by more than ten percentage points over the coming months. Though I'll write here if I do.

My prediction: 70% chance that by August 2029 there will be a large reduction in investment in AI and a corresponding market crash in AI company stocks, etc, and that both will continue to be for at least three months.

 

For:

  • Large model labs losing money
    • OpenAI made loss of ~$5 billio
... (read more)

Update: back up to 60% chance. 

I overreacted before IMO on the updating down to 40% (and undercompensated when updating down to 80%, which I soon after thought should have been 70%).

The leader in turns of large model revenue, OpenAI has basically failed to build something worth calling GPT-5, and Microsoft is now developing more models in-house to compete with them. If OpenAI fails on the effort to combine its existing models into something new and special (likely), that’s a blow to perception of the industry.

A recession might also be coming this year, or at least in the next four years, which I made a prediction about before.

2
Remmelt
Update: back up to 70% chance. Just spent two hours compiling different contributing factors. Now I weighed those factors up more comprehensively, I don't expect to change my prediction by more than ten percentage points over the coming months. Though I'll write here if I do. My prediction: 70% chance that by August 2029 there will be a large reduction in investment in AI and a corresponding market crash in AI company stocks, etc, and that both will continue to be for at least three months.   For: * Large model labs losing money * OpenAI made loss of ~$5 billion last year. * Takes most of the consumer and enterprise revenue, but still only $3.7 billion. * GPT 4.5 model is the result of 18 months of R&D, but only a marginal improvement in output quality, while even more compute intensive. * If OpenAI publicly fails, as the supposed industry leader, this can undermine the investment narrative of AI as a rapidly improving and profitable technology, and trigger a market meltdown. * Commoditisation * Other models by Meta, etc, around as useful for consumers. * DeepSeek undercuts US-designed models with compute-efficient open-weights alternative. * Data center overinvestment * Microsoft cut at least 14% of planned data center expansion. * Subdued commercial investment interest. * Some investment firm analysts skeptical, and second-largest VC firm Sequoia Capital also made a case of lack of returns for the scale of investment ($600+ billion). * SoftBank is the main other backer of the Stargate data center expansion project, and needs to raise debt to do raise ~$18 billion. OpenAI also needs to raise more investment funds next round to cover ~$18 billion, with question whether there is interest * Uncertainty US government funding * Mismatch between US Defense interest and what large model labs are currently developing. * Model 'hallucinations' get in the way of deployment of LLMs on the battlefield, given reliabil

Update: back up to 50% chance. 

Noting Microsoft’s cancelling of data center deals. And the fact the ‘AGI’ labs are still losing cash, and with DeepSeek are competing increasingly on a commodity product. 

2
Remmelt
Update: back up to 60% chance.  I overreacted before IMO on the updating down to 40% (and undercompensated when updating down to 80%, which I soon after thought should have been 70%). The leader in turns of large model revenue, OpenAI has basically failed to build something worth calling GPT-5, and Microsoft is now developing more models in-house to compete with them. If OpenAI fails on the effort to combine its existing models into something new and special (likely), that’s a blow to perception of the industry. A recession might also be coming this year, or at least in the next four years, which I made a prediction about before.

Update: 40% chance. 

I very much underestimated/missed the speed of tech leaders influencing the US government through the Trump election/presidency. Got caught flat-footed by this. 

I still think it’s not unlikely for there to be an AI crash as described above within the next 4 years and 8 months but it could be from levels of investment much higher than where we are now. A “large reduction in investment” at that level looks a lot different than a large reduction in investment from the level that markets were at 4 months ago. 

4
Remmelt
Update: back up to 50% chance.  Noting Microsoft’s cancelling of data center deals. And the fact the ‘AGI’ labs are still losing cash, and with DeepSeek are competing increasingly on a commodity product. 

We ended up having a private exchange about it. 

Basically, organisers spend more than half of their time on general communications and logistics to support participants get to work. 

And earmarking stipends to particular areas of work seems rather burdensome administratively, though I wouldn’t be entirely against it if it means we can cover more people’s stipends.

Overall, I think we tended not to allow differentiated fundraising before because it can promote internal conflicts, rather than having people come together to make the camp great.

Answer by Remmelt3
0
0

Here's how I specify terms in the claim:

  • AGI is a set of artificial components, connected physically and/or by information signals over time, to in aggregate sense and act autonomously over many domains.
    • 'artificial' as configured out of a (hard) substrate that can be standardised to process inputs into outputs consistently (vs. what our organic parts can do).
    • 'autonomously' as continuing to operate without needing humans (or any other species that share a common ancestor with humans).
  • Alignment is at the minimum the control of the AGI's components (as modifie
... (read more)

Update: reverting my forecast back to 80% chance likelihood for these reasons.

3
Remmelt
Update: 40% chance.  I very much underestimated/missed the speed of tech leaders influencing the US government through the Trump election/presidency. Got caught flat-footed by this.  I still think it’s not unlikely for there to be an AI crash as described above within the next 4 years and 8 months but it could be from levels of investment much higher than where we are now. A “large reduction in investment” at that level looks a lot different than a large reduction in investment from the level that markets were at 4 months ago. 

I'm also feeling less "optimistic" about an AI crash given:

  1. The election result involving a bunch of tech investors and execs pushing for influence through Trump's campaign (with a stated intention to deregulate tech).
  2. A military veteran saying that the military could be holding up the AI industry like "Atlas holding the globe", and an AI PhD saying that hyperscaled data centers, deep learning, etc, could be super useful for war.

I will revise my previous forecast back to 80%+ chance.

Just found a podcast on OpenAI’s bad financial situation.

It’s hosted by someone in AI Safety (Jacob Haimes) and an AI post-doc (Igor Krawzcuk).

https://kairos.fm/posts/muckraiker-episodes/muckraiker-episode-004/

There are bunch of crucial considerations here. I’m afraid it would take too much time to unpack those.

Happy though to have had this chat!

As a 1st approximation, I assume humans will be selecting AIs which benefit them, not AIs which maximally increase economic growth.

The problem here is that AI corporations are increasingly making decisions for us. 
See this chapter.

Corporations produce and market products to increase profit (including by replacing their fussy expensive human parts with cheaper faster machines that do good-enough work.)

To do that they have to promise buyers some benefits, but they can also manage to sell products by hiding the negative externalities. See cases Big Tobacco, Big Oil, etc.

2
Vasco Grilo🔸
I agree it makes sense to model corporations as maximising profit, to a 1st approximation. However, since humans ultimately want to be happy, not increasing gross world product, I assume people will tend to pay more for AIs which are optimising for human welfare instead of economic growth. So I expect corporations developping AIs optimising for something closer to human welfare will be more successful/profitable than ones developping AIs which maximally increase economic growth. That being said, if economic growth refers to the growth of the human economy (instead of the growth of the AI economy too), I guess optimising for economic growth will lead to better outcomes for humans, because this has historically been the case.

I am open to a bet similar to this one.

I would bet on both, on your side.
 

Potentially relatedly, I think massive increases in unemployment are very unlikely.

I see you cite statistics of previous unemployment rates as an outside view, compensating against the inside view. Did you look into the underlying rate of job automation? I'd be curious about that. If that underlying rate has been trending up over time, then there is a concern that at some point the gap might not be filled with re-employment opportunities.

AI Safety inside views are wrong for vari... (read more)

2
Vasco Grilo🔸
Thanks for clarifying! Fair! I did not look into that. However, the rate of automation (not the share of automated tasks) is linked to economic growth, and this used to be much lower in the past. According to Table 1 (2) of Hanson 2000, the global economy used to double once every 230 k (224 k) years in hunting and gathering period of human history. Today it doubles once every 20 years or so[1]. Despite a much higher growth rate, and therefore a way higher rate of automation, the unemployment rate is still relatively low (5.3 % globally in 2022). So I still think it is very unlikely that faster automation in the next few years would lead to massive unemployment. Longer term, over decades to centuries, I can see AI coming to perform the vast majority of economically valuable tasks. However, I believe humans will only allow this to happen if they get to benefit. As a 1st approximation, I assume humans will be selecting AIs which benefit them, not AIs which maximally increase economic growth. 1. ^ The doubling time for 3 % annual growth is 23.4 years (= LN(2)/LN(1.03)).

Donation opportunities for restricting AI companies:

... (read more)

Hey, my apologies for taking even longer to reply (had family responsibilities this month). 

I will read that article on why Chernobyl-style events are not possible with modern reactors. Respecting you for the amount of background research you must have done in this area, and would like to learn more.

Although I think the probability of human extinction over the next 10 years is lower than 10^-6.

You and I actually agree on this with respect to AI developments. I don’t think the narratives I read of a large model recursively self-improving itself internally make sense.

I wrote a book for educated laypeople explaining how AI corporations would cause increasing harms leading eventually  to machine destruction of our society and ecosystem.

Curious for your own thoughts here. 

4
Vasco Grilo🔸
Thanks for sharing! You may want to publish a post with a summary of the book. Potentially relatedly, I think massive increases in unemployment are very unlikely. If you or anyone you know are into bets, and guess the unemployment rate in the United States will reach tens of % in the next few years, I am open to a bet similar to this one.

Basically I'm upvoting what you're doing here, which I think is more important than the text itself.

Thanks for recognising the importance of doing the work itself. We are still scrappy so we'll find ways to improve over time.
 

especially that you should have run this past a bunch of media savvy people before releasing

If you know anyone with media experience who might be interested to review future drafts, please let me know. 

I agree we need to improve on our messaging.

 

This is great!  Appreciating your nitty-gritty considerations  

(1) There's a good chance the outcome will be some form of "catch and release" -- it's usually easier to deal with isolated protestors who do not cause violence or significantly damage property in this manner rather than by pursuing criminal charges to trial.

“Catch and release” is what’s happening right now. However, as we keep repeating the barricades, and as hopefully more and more protestors join us, it would highly surprise me if the police and court system just allow us to keep b... (read more)

3
Jason
Keeping it at a general information level -- There are a few main ways that judges "make law" when rendering judicial decisions. One involves creating precedent that (loosely) binds their court and (more strictly) binds certain lower courts.[1] Note that, in most systems, this only occurs when the court designates its opinion as precedential ("published").[2] Moreover, only the "holding" (~what was necessary to decide the case) rather than the "dicta" (anything else) become precedent. For better or worse, the courts do not clearly label the difference. Another method of impact involves the binding effect of the court's judgment on the parties. In the right procedural posture, this can be very powerful against a government defendant. That usually happens when the only way to provide relief for the plaintiff's legally recognized injuries in a suit is to award sweeping relief against the government that is within the court's power.  In other situations, it can be largely useless -- there are circumstances in which so-called collateral estoppel (ye olden name) or issue preclusion (the modern name) won't run against the government when it would against a private litigant. This makes a lot of sense in the criminal context because the government cannot appeal an acquittal, and we generally don't want to bind a party too much if it lacks the ability to appeal. People talk about "precedent" in a looser sense -- you can point to Judge Smith's decision in a prior case and try to convince Judge Jones that he should rule the same way. Also, the prosecutor might see Judge Smith's decision and decide not to file a similar case in the future. Seeking this sort of "precedent" is a valid activist strategy, but it's important to recognize when it is more or less likely to work. This kind of "precedent" is more likely to be effective when there is limited on-point binding precedent. It can also be effective in causing an opponent to update the odds that litigation won't go well fo

Thanks for the kind words!

I personally think it would be helpful to put more emphasis on how OpenAI’s reckless scaling and releases of models is already concretely harming ordinary folks (even though no major single accident has shown up yet).

Eg.

  • training on personal/copyrighted data
  • job losses because of the shoddy replacement of creative workers (and how badly OpenAI has treated workers it paid)
  • school ‘plagiarism’, disinformation, and deepfakes
  • environmental harms of scaling compute.  

Thank you for the specific feedback on the press release!  I broadly agree with it, and I think it’s going to be useful for improving future texts.

Yes, thank you for mentioning this  

I made a mistake in checking that number. 

See also comment here

To clarify for future reference, I do think it’s likely (80%+) that at some point over the next 5 years there will be a large reduction in investment in AI and a corresponding market crash in AI company stocks, etc, and that both will continue to be for at least three months.

Update: I now think this is 90%+ likely to happen (from original prediction date).

2
Remmelt
Update: reverting my forecast back to 80% chance likelihood for these reasons.

But it’s weird that I cannot find even a good written summary of Bret’s argument online (I do see lots of political podcasts).

I found an earlier scenario written by Bret that covers just one nuclear power plant failing and that does not discuss the risk of a weakening magnetic field.

The quotes from the OECD Nuclear Energy Agency’s report were interesting.

moving nuclear fuel stored in pools into dry casket storage The extent to which we can do this is limited because spent fuel must be stored for one to ten years in spent fuel pools while the shorter-lived isotopes decay before it's ready to be moved to dry cask storage.

I did not know this. I added an edit to the post: “nuclear waste already stored in pools for 5 years”.

I don't think an environmental radioisotope release can realistically give people across the world acute radiat

... (read more)
1
ABlank
I don't have anything to cite, but everything i've read about real or hypothetical major nuclear accidents affecting large areas talks about harms like increased cancer risk over the course of years. Dying within days or weeks as described in the original post requires orders of magnitude higher doses of radiation in a shorter time. I don't think it's possible to get those kinds of doses from the amount of radioactive material that can realistically be dispersed kilometers away from the accident. (Being concerned about loss of safely usable land for living and farming is reasonable and i'm only complaining about this point because you specifically described acute radiation syndrome.)

Thanks, looking forward to reading your thoughts!

Regarding 1., I would value someone who has researched this give more insight into:

A. How long diesel generators could be expected to be supplied with diesel when there is some continental electricity outage of a year (or longer). This is hard to judge. My intuition is that society would be in chaos and that maintaining diesel supplies would be extremely tough to manage.

B. What is minimally required in the long process of shutting down a nuclear power plant? Including but not limited to diesel or other backup generator supplies.

Regarding 2., I do not see h... (read more)

4
jackva
I'll address the rest later (hopefully this weekend), but just to clarify that I did not mean to call you or your summary conspiratorial, Remmelt! I only meant the substantive claim and coming from Bret Weinstein, but I'll address this with more time and the rest of your comment later.

Fixed it!  You can use either link now to share with your friends.

Igor Krawzcuk, an AI PhD researcher, just shared more specific predictions:

“I agree with ed that the next months are critical, and that the biggest players need to deliver. I think it will need to be plausible progress towards reasoning, as in planning, as in the type of stuff Prolog, SAT/SMT solvers etc. do.

I'm 80% certain that this literally can't be done efficiently with current LLM/RL techniques (last I looked at neural comb-opt vs solvers, it was _bad_), the only hope being the kitchen sink of scale, foundation models, solvers _and_ RL

If OpenAI/Anthr... (read more)

To clarify for future reference, I do think it’s likely (80%+) that at some point over the next 5 years there will be a large reduction in investment in AI and a corresponding market crash in AI company stocks, etc, and that both will continue to be for at least three months.

Ie. I think we are heading for an AI winter. 
It is not sustainable for the industry to invest 600+ billion dollars per year in infrastructure and teams in return for relatively little revenue and no resulting profit for major AI labs.

At the same time, I think that within the next ... (read more)

2
Remmelt
Update: I now think this is 90%+ likely to happen (from original prediction date).
2
Remmelt
Igor Krawzcuk, an AI PhD researcher, just shared more specific predictions: “I agree with ed that the next months are critical, and that the biggest players need to deliver. I think it will need to be plausible progress towards reasoning, as in planning, as in the type of stuff Prolog, SAT/SMT solvers etc. do. I'm 80% certain that this literally can't be done efficiently with current LLM/RL techniques (last I looked at neural comb-opt vs solvers, it was _bad_), the only hope being the kitchen sink of scale, foundation models, solvers _and_ RL … If OpenAI/Anthropic/DeepMind can't deliver on promises of reasoning and planning (Q*, Strawberry, AlphaCode/AlphaProof etc.) in the coming months, or if they try to polish more turds into gold (e.g., coming out with GPT-Reasoner, but only for specific business domains) over the next year, then I would be surprised to see the investments last to make it happen in this AI summer.” https://x.com/TheGermanPole/status/1826179777452994657

What are you thinking about in terms of pre-harm enforcement? 

I’m thinking about advising premarket approval – a requirement to scope model designs around prespecified uses and having independent auditors vet the safety tests and assessments.

The report is focussed on preventing harms of technology to people using or affected by that tech.

It uses FDA’s mandate of premarket approval and other processes as examples of what could be used for AI.

Restrictions to economic productivity and innovation is a fair point of discussion. I have my own views on this – generally I think the negative assymetry around new scalable products being able to do massive harm gets neglected by the market. I’m glad the FDA exists to counteract that.

The FDA’s slow response to ramping up COVID vaccines during the pandemic... (read more)

They mentioned that line at the top of the 80k Job board.

Still do I see.

“Handpicked to help you tackle the world's most pressing problems with your career.”

https://jobs.80000hours.org/

Ah, I wasn't clear. To bet that AI will not kill us all by the end of 2027. 

I don't think that makes sense, given the world-complexity "AI" would need to learn and evolve and get tinkered to be able to navigate. I've had some conversations with Greg about this.

Load more