Hide table of contents
This is a linkpost for https://youtu.be/VcVfceTsD0A

Context: Max Tegmark is a cofounder of the Future of Life Institute and a leading AI Safety researcher at MIT. He recorded a podcast a few hours before the Future of Life Institute published a call to halt training on large-scale AI models more sophisticated than GPT-4 for six months. This is a summary of that podcast.

Summary

  • Max Tegmark gives arguments for why we'll develop Artificial General Intelligence much faster than expected by most AI safety researchers. Arguments include how: 
    • Training models on simple tasks (like predicting the next word in a sentence) trains surprisingly large complex/general behaviour. 
    • Current models are known to be inefficient and tiny improvements seem to be having outsized gains (not explained by more computational resources/data alone)
  • Since he doesn't see the rate of AI safety research matching the rate of AI development, he sees a coordinated global pause as the only chance to create necessary changes. 
  • Research questions he prioritises include:
    • How do we create AI models which can output formal mathematical proofs about their workings to check if they're aligned?
    • How do we extract and analyse an AI model's functions from its parameters?
    • How do we train an AI model to be hesitant and seek feedback before executing actions?

 

Raw Notes

  • AI interpretability research is revealing a lot of inefficiencies in the way modern models like transformers work. Ex Researchers see that AI models will store data like facts (“Where is the Eiffel Tower?”) in sparse matrices in their weights and this could be made a lot more efficient. If researchers add tiny fixes to get around this, there could be exponential improvements in performance instead of just incremental ones due to more data/computational resources. 
    • A good analogy that Max Tegmark gave for this was about creating the ability to fly. To mimic how birds fly, it took us until modern robotics today. But we invented planes to fly and other flying machines like hot air balloons hundreds of years before that. This was possible due to simpler and less efficient designs - even today’s planes are less energy-efficient than birds. 
    • The generalisable point is that we don’t need to mimic nature’s highly-optimised capabilities to achieve the end result more quickly and less efficiently. To create human-level AI, you don’t need to figure out how brains work. To get GPT-4 level reasoning ability, you can just create an inefficient model that’s only trying to predict which words come next in a sentence.
  • He believes it’s possible to slow down the AI arms race if we have enough public pressure so that all countries realise that an out-of-control model is bad for everyone. He says rapidly building AI systems that aren’t fully safe “isn’t an arms race, it’s a suicide race.
  • Currently, we’re violating all the reasonable safety norms with AI development. 
    • We have arms race dynamics between companies and states. 
    • People are teaching AI systems to code (which researchers recommended avoiding so that AI systems aren’t able to create other software). 
    • People have already taught AI systems about manipulating human psychology (every recommender system on a major social media platform knows us better than other humans know us). 
    • People are giving the most advanced AI systems access to the Internet. RE: Bing Chat, AutoGPT, …
    • People are letting the most advanced AI systems interact with each other and improve each other. RE: AutoGPT prompting itself, using vector databases to hack together long-term memory for AI systems.
  • We’re already seeing difficulties in regulating AI because the lobbyists for the largest tech companies are more so controlling policy than the policymakers are controlling the tech companies. Ex: In the EU, the AI Act kept getting pushback from individual companies to exclude their models from the act. 
    • And it’s entirely possible that the most wealthy companies keep consolidating even more because of compute/data requirements creating economies of scale.
  • He mentions how a colleague of his created a mathematical proof in general that continuing to optimise one goal forever leads you to make things worse eventually. Formalisation of what many folks might claim via their everyday experience.
  • He says that throughout history, groups of people (and other animals) that were no longer needed usually were treated very badly. He gives an example of how there were many horses before cars, but they were largely cast aside when they were no longer needed. He doesn’t give a specific example for humans. 
  • Right now, he says that we’ve started to automate jobs that people really love, not just jobs that are boring, dangerous, or undesirable. Ex: A lot of artists right now are no longer needed because AI models can generate photos, videos, 3D assets, etc.
  • He says that optimising AI to be truthful is very largely useful. He doesn’t elaborate much, but does give one example of how there will be fewer misinformation problems and ensuing conflicts if every human could trust the output of some AI model in a guaranteed way. Ie. This would get around the need to rely on some central authority to be truthful.
    • Technically, you could optimise truthfulness by using Brier loss evaluation metrics. 
  • He says that it’s a lot easier to verify if a mathematical proof is correct than to generate a mathematical proof. He wants to build on this foundation to create reliable checkers for advanced AI algorithms to see if they’re “trustworthy.” Then, discarding the algorithms that turn out not be trustworthy. 
    • Right now, this idea seems like it’s too vague to be useful. 
    • Another idea that’s currently at the vague intuition level is trying to create systems that extract out functioning mechanisms from a model’s parameters. Then, making those mechanisms run in an auditable way. 
  • He thinks we’re already past the point where cutting-edge systems like GPT-4 can be made open-source. Now, he thinks the potential for harm is just too high compared to the benefits of having multiple technical eyes to fix issues.
    • He sees that current safety checks on these systems try to prevent them from spreading misinformation, spreading harmful information, and conducting cyberattacks. He thinks these are good problems to work on, but they don’t have the same scale of importance as the potential of an advanced AI system getting full-blown general capabilities or the damage that already-general systems can do to the economy. 
    • He’s not concerned about AI systems intentionally trying to cause harm to humans relative to the probability of them accidentally doing it or doing it as a side action to get to another goal. 
  • Never mind changing what the education system is teaching kids, we need to change the education system so that it doesn’t ever define what kids should be taught during over a decade of change. 
  • The largest cost of nuclear weapons would be the catastrophe it causes on food systems. He says this recent model estimates 98-99% of people in northern countries dying due to famines following a global nuclear war.
Comments4


Sorted by Click to highlight new comments since:

Madhav - thanks for a very helpful summary of Max Tegmark's remarks. I agree with most of your comments about his views.

His line that 'this isn’t an arms race, it’s a suicide race' seems pretty compelling as a counter-argument against the view -- very commonly expressed on Twitter -- that the US must 'push ahead with AI, so we don't fall behind China'. 

Just listened to a podcast interview of yours, Geoffrey Miller (Manifold, with Steve Hsu). Do you really believe that it is viable to impose a very long pause (you mention 'just a few centuries'). The likelihood of such a thing taking place seems to me more than extremely remote -at least until we get a pragmatic example of the harm AI can do, a Trinity test of sorts.

Manuel - it may not be likely that we can impose a very long pause (on the order of centuries).

My main goal in proposing that was to remind people that with AI, we're talking about a 'major evolutionary transition' comparable to the evolution of multicellular life, or the evolution of human intelligence. Normally these take place on the time-span of hundreds of thousands of years to tens of millions of years. 

If AI development holds all the potential upside that we hope for, but it also threatens some of the downside risks that we dread, it may be useful to be thinking on these evolutionary time scales. Rather than the 'quarterly profit' schedules that many Big Tech companies think about.

Does anyone one have the time stamp, when he argues that it's easier to show that a mathematical proof is correct than to find the proof. I am working on making that argument rigorous for the case of an AI planer and would like to reference the conversation.

Edit: I found it: it's at 01:46:50

Curated and popular this week
 ·  · 16m read
 · 
Applications are currently open for the next cohort of AIM's Charity Entrepreneurship Incubation Program in August 2025. We've just published our in-depth research reports on the new ideas for charities we're recommending for people to launch through the program. This article provides an introduction to each idea, and a link to the full report. You can learn more about these ideas in our upcoming Q&A with Morgan Fairless, AIM's Director of Research, on February 26th.   Advocacy for used lead-acid battery recycling legislation Full report: https://www.charityentrepreneurship.com/reports/lead-battery-recycling-advocacy    Description Lead-acid batteries are widely used across industries, particularly in the automotive sector. While recycling these batteries is essential because the lead inside them can be recovered and reused, it is also a major source of lead exposure—a significant environmental health hazard. Lead exposure can cause severe cardiovascular and cognitive development issues, among other health problems.   The risk is especially high when used-lead acid batteries (ULABs) are processed at informal sites with inadequate health and environmental protections. At these sites, lead from the batteries is often released into the air, soil, and water, exposing nearby populations through inhalation and ingestion. Though data remain scarce, we estimate that ULAB recycling accounts for 5–30% of total global lead exposure. This report explores the potential of launching a new charity focused on advocating for stronger ULAB recycling policies in low- and middle-income countries (LMICs). The primary goal of these policies would be to transition the sector from informal, high-pollution recycling to formal, regulated recycling. Policies may also improve environmental and safety standards within the formal sector to further reduce pollution and exposure risks.   Counterfactual impact Cost-effectiveness analysis: We estimate that this charity could generate abou
sawyer🔸
 ·  · 2m read
 · 
Note: This started as a quick take, but it got too long so I made it a full post. It's still kind of a rant; a stronger post would include sources and would have gotten feedback from people more knowledgeable than I. But in the spirit of Draft Amnesty Week, I'm writing this in one sitting and smashing that Submit button. Many people continue to refer to companies like OpenAI, Anthropic, and Google DeepMind as "frontier AI labs". I think we should drop "labs" entirely when discussing these companies, calling them "AI companies"[1] instead. While these companies may have once been primarily research laboratories, they are no longer so. Continuing to call them labs makes them sound like harmless groups focused on pushing the frontier of human knowledge, when in reality they are profit-seeking corporations focused on building products and capturing value in the marketplace. Laboratories do not directly publish software products that attract hundreds of millions of users and billions in revenue. Laboratories do not hire armies of lobbyists to control the regulation of their work. Laboratories do not compete for tens of billions in external investments or announce many-billion-dollar capital expenditures in partnership with governments both foreign and domestic. People call these companies labs due to some combination of marketing and historical accident. To my knowledge no one ever called Facebook, Amazon, Apple, or Netflix "labs", despite each of them employing many researchers and pushing a lot of genuine innovation in many fields of technology. To be clear, there are labs inside many AI companies, especially the big ones mentioned above. There are groups of researchers doing research at the cutting edge of various fields of knowledge, in AI capabilities, safety, governance, etc. Many individuals (perhaps some readers of this very post!) would be correct in saying they work at a lab inside a frontier AI company. It's just not the case that any of these companies as
Dorothy M.
 ·  · 5m read
 · 
If you don’t typically engage with politics/government, this is the time to do so. If you are American and/or based in the U.S., reaching out to lawmakers, supporting organizations that are mobilizing on this issue, and helping amplify the urgency of this crisis can make a difference. Why this matters: 1. Millions of lives are at stake 2. Decades of progress, and prior investment, in global health and wellbeing are at risk 3. Government funding multiplies the impact of philanthropy Where things stand today (February 27, 2025) The Trump Administration’s foreign aid freeze has taken a catastrophic turn: rather than complying with a court order to restart paused funding, they have chosen to terminate more than 90% of all USAID grants and contracts. This stunningly reckless decision comes just 30 days into a supposed 90-day review of foreign aid. This will cause a devastating loss of life. Even beyond the immediate deaths, the long-term consequences are dire. Many of these programs rely on supply chains, health worker training, and community trust that have taken years to build, and which have already been harmed by U.S. actions in recent weeks. Further disruptions will actively unravel decades of health infrastructure development in low-income countries. While some funding may theoretically remain available, the reality is grim: the main USAID payment system remains offline and most staff capable of restarting programs have been laid off. Many people don’t believe these terminations were carried out legally. But NGOs and implementing partners are on the brink of bankruptcy and insolvency because the government has not paid them for work completed months ago and is withholding funding for ongoing work (including not transferring funds and not giving access to drawdowns of lines of credit, as is typical for some awards). We are facing a sweeping and permanent shutdown of many of the most cost-effective global health and development programs in existence that sa