Hide table of contents

We, A Happier World, just uploaded a video on value lock-in, inspired by Will MacAskill's book What We Owe The Future! 

This is part of a whole series we're making on the book, full playlist here.

Thanks to Sarah Emminghaus for her help with the script!

Transcript

Sources are marked with an asterisk. Text might differ slightly in wording from the final video.

Hundred schools of thought

2600 years ago, China went through a long period of conflict that is now known as the Warring states era. But it also brought about a time with many philosophical and cultural experiments that is now known as the hundred schools of thought. That’s when Confucianism was born – the philosophy of Kong Fuzi who believed that self-improvement led to a spiritual transformation. Confucianism encouraged respect for your parents and obedience to authority, rulers and the state. The ethics depended on relationships between people rather than the actions themselves: A son beating their father is not okay, but the opposite is.**

There were a few other popular philosophies at the time; for example legalism. Legalists were strong proponents of heavy punishments for wrongdoings, a powerful military and a strong state, they believed people were selfish and needed heavy guidance.

Then there were the mohists – at the time, they were the confucianists’ main rival. Mohists believed that we should care about other people as much as we care about ourselves. And that we should take whatever actions benefited the most people. They proposed owning no luxury and consuming less.

The rivalism ended in 221 BC when the legalism-influenced Qin conquered China and took strong measures against all competing schools of thought – apparently legalism had won. That all changed when the dynasty ended just 15 years later and Confucianism turned out to be the new popular ideology.* Since then, all Chinese dynasties embraced Confucianism until the Qing dynasty ended in 1912. When Mao and the communist party started ruling China in 1949 it got suppressed, but it remained popular and it’s being revived today.

The popularity of Confucianism is a great example of value lock-in: a situation where one set of values wins against others and stays in place for a very long time.

Other examples include Christianity and Islam: The bible and the Quran are still the best-selling books today!

Risks of locking in current values

In general when we look at values from the past – be it 10, 50 or 200 years ago – it feels like we progressed towards the better. There’s no way we would want those “outdated” values to still persist. But what makes a lot of us so sure that our current values are good? They might be better in some ways, but just like we do now, people in the future will probably look at our current values and find a lot of them abhorrent. They will probably be glad we overcame those “outdated” values.

Right now a lot of pretty different worldviews are competing against each other – similar as in the time of the hundred schools of thought in China. No one school of thought has won – yet. 

Now what is interesting about this with regards to the long term future: How does a school of thought win against others? How do values get locked in? What values from today do we want to lock in – if any? Which ones are maybe already locked in even though they might not be what is best for humanity?

This video is part of a series based on Oxford philosopher Will MacAskill’s new book called What We Owe The Future. The book makes the case for caring about our longterm future and explores what we can do to have a positive impact on it. In a previous episode we told the story of Benjamin Lay, an inspiring abolitionist. In another one we argued why caring about our longterm future is important.

Will MacAskill mentions three ways values could get locked in in the future: If we set up a world government, through space colonisation, and through the development of powerful AI. The latter is perhaps the scariest. So far, Artificial Intelligence is narrow – meaning, machines only know how to perform narrow tasks, like beating us at chess. No AI can actually fully replicate human intelligence over all possible tasks. When AI exceeds human capabilities across all domains, and that might happen at some point sooner rather than later, there’s the massive risk values would be locked in that are not what is best for humanity. Perhaps by the AI somehow imposing them on us. 

This is already happening to some degree today through social media algorithms. 

We know this topic is incredibly broad and complex. We’re also not sure whether this really would happen. After all, it sounds a bit ridiculous. We will talk more about AI in our next video on existential threats, so subscribe and ring that notification bell to get notified when it comes out!

So how do we keep the wrong values from being locked in? 

In general: We should try to keep our minds as open as possible for as long as possible. As such, we should promote free speech and create a marketplace of ideas wherever possible. Whenever multiple schools of thought are encouraged, it gets less likely for one of them to win and be locked in.

One idea we like a lot is the idea of ideological charter cities: autonomous communities with their own laws that try out different ideas as an experiment – to see how well some theories translate into the real world without the specific policies being difficult to reverse. One real life example again happened in China not that long ago. In 1979 a special economic zone was created around the city of Shenzhen – with more liberal economic policies than in the rest of the country. The average yearly income there grew by a factor of two hundred over forty years, the experiment inspired broader reforms across the country. Since then, hundreds of millions of people in China have gotten out of poverty.

The idea of an ideological charter city could also be tried out by marxists, environmentalists or anarchists! That way we could see which ideas work best in the real world and learn something from it.

Another thing that can help is to have more open borders. If people are allowed to migrate more easily, people can vote with their feet which countries’ values they prefer.

Then there are values that are already helping us to get to morally better societies. Reason, reflection and empathy are probably among them. Engaging in good faith arguments, being open to other people’s viewpoints, empathising with people not in our inner circle – all those practices can help getting to an improved point of view and better morals.

There’s a paradox here: Making sure values don’t get locked in means locking in values like these. But this is a paradox we’re happy to live with.

Conclusion

If you thought this was interesting, you should definitely check out Will’s book. In the next episode we’ll be talking about the possible end of humanity. So don’t forget to subscribe!

Comments2


Sorted by Click to highlight new comments since:

Love this - love the set design, love that you've come on camera, great script and interesting topic! Everything is awesome! 

Curated and popular this week
 ·  · 38m read
 · 
In recent months, the CEOs of leading AI companies have grown increasingly confident about rapid progress: * OpenAI's Sam Altman: Shifted from saying in November "the rate of progress continues" to declaring in January "we are now confident we know how to build AGI" * Anthropic's Dario Amodei: Stated in January "I'm more confident than I've ever been that we're close to powerful capabilities... in the next 2-3 years" * Google DeepMind's Demis Hassabis: Changed from "as soon as 10 years" in autumn to "probably three to five years away" by January. What explains the shift? Is it just hype? Or could we really have Artificial General Intelligence (AGI)[1] by 2028? In this article, I look at what's driven recent progress, estimate how far those drivers can continue, and explain why they're likely to continue for at least four more years. In particular, while in 2024 progress in LLM chatbots seemed to slow, a new approach started to work: teaching the models to reason using reinforcement learning. In just a year, this let them surpass human PhDs at answering difficult scientific reasoning questions, and achieve expert-level performance on one-hour coding tasks. We don't know how capable AGI will become, but extrapolating the recent rate of progress suggests that, by 2028, we could reach AI models with beyond-human reasoning abilities, expert-level knowledge in every domain, and that can autonomously complete multi-week projects, and progress would likely continue from there.  On this set of software engineering & computer use tasks, in 2020 AI was only able to do tasks that would typically take a human expert a couple of seconds. By 2024, that had risen to almost an hour. If the trend continues, by 2028 it'll reach several weeks.  No longer mere chatbots, these 'agent' models might soon satisfy many people's definitions of AGI — roughly, AI systems that match human performance at most knowledge work (see definition in footnote). This means that, while the compa
 ·  · 4m read
 · 
SUMMARY:  ALLFED is launching an emergency appeal on the EA Forum due to a serious funding shortfall. Without new support, ALLFED will be forced to cut half our budget in the coming months, drastically reducing our capacity to help build global food system resilience for catastrophic scenarios like nuclear winter, a severe pandemic, or infrastructure breakdown. ALLFED is seeking $800,000 over the course of 2025 to sustain its team, continue policy-relevant research, and move forward with pilot projects that could save lives in a catastrophe. As funding priorities shift toward AI safety, we believe resilient food solutions remain a highly cost-effective way to protect the future. If you’re able to support or share this appeal, please visit allfed.info/donate. Donate to ALLFED FULL ARTICLE: I (David Denkenberger) am writing alongside two of my team-mates, as ALLFED’s co-founder, to ask for your support. This is the first time in Alliance to Feed the Earth in Disaster’s (ALLFED’s) 8 year existence that we have reached out on the EA Forum with a direct funding appeal outside of Marginal Funding Week/our annual updates. I am doing so because ALLFED’s funding situation is serious, and because so much of ALLFED’s progress to date has been made possible through the support, feedback, and collaboration of the EA community.  Read our funding appeal At ALLFED, we are deeply grateful to all our supporters, including the Survival and Flourishing Fund, which has provided the majority of our funding for years. At the end of 2024, we learned we would be receiving far less support than expected due to a shift in SFF’s strategic priorities toward AI safety. Without additional funding, ALLFED will need to shrink. I believe the marginal cost effectiveness for improving the future and saving lives of resilience is competitive with AI Safety, even if timelines are short, because of potential AI-induced catastrophes. That is why we are asking people to donate to this emergency appeal
 ·  · 1m read
 · 
We’ve written a new report on the threat of AI-enabled coups.  I think this is a very serious risk – comparable in importance to AI takeover but much more neglected.  In fact, AI-enabled coups and AI takeover have pretty similar threat models. To see this, here’s a very basic threat model for AI takeover: 1. Humanity develops superhuman AI 2. Superhuman AI is misaligned and power-seeking 3. Superhuman AI seizes power for itself And now here’s a closely analogous threat model for AI-enabled coups: 1. Humanity develops superhuman AI 2. Superhuman AI is controlled by a small group 3. Superhuman AI seizes power for the small group While the report focuses on the risk that someone seizes power over a country, I think that similar dynamics could allow someone to take over the world. In fact, if someone wanted to take over the world, their best strategy might well be to first stage an AI-enabled coup in the United States (or whichever country leads on superhuman AI), and then go from there to world domination. A single person taking over the world would be really bad. I’ve previously argued that it might even be worse than AI takeover. [1] The concrete threat models for AI-enabled coups that we discuss largely translate like-for-like over to the risk of AI takeover.[2] Similarly, there’s a lot of overlap in the mitigations that help with AI-enabled coups and AI takeover risk — e.g. alignment audits to ensure no human has made AI secretly loyal to them, transparency about AI capabilities, monitoring AI activities for suspicious behaviour, and infosecurity to prevent insiders from tampering with training.  If the world won't slow down AI development based on AI takeover risk (e.g. because there’s isn’t strong evidence for misalignment), then advocating for a slow down based on the risk of AI-enabled coups might be more convincing and achieve many of the same goals.  I really want to encourage readers — especially those at labs or governments — to do something
Relevant opportunities
60
· · 3m read