Hide table of contents

Epistemic status: highly certain, or something

The Spending What We Must 💸11% pledge 

In short: Members pledge to spend at least 11% of their income on effectively increasing their own productivity.

This pledge is likely higher-impact for most people than the Giving What We Can 🔸10% Pledge, and we also think the name accurately reflects the non-supererogatory moral beliefs of many in the EA community.

Example

Charlie is a software engineer for the Centre for Effective Future Research. Since Charlie has taken the SWWM 💸11% pledge, rather than splurge on a vacation, they decide to buy an expensive noise-canceling headset before their next EAG, allowing them to get slightly more sleep and have 104 one-on-one meetings instead of just 101. In one of the extra three meetings, they chat with Diana, who is starting an AI-for-worrying-about-AI company, and decide to become a cofounder. The company becomes wildly successful, and Charlie's equity share allows them to further increase their productivity to the point of diminishing marginal returns, then donate $50 billion to SWWM.

The 💸💸💸 Badge

If you've taken the SWWM 💸11% Pledge, we'd appreciate if you could add three 💸💸💸 "stacks of money with wings" emoji to your social media profiles. We chose three emoji because we think the 💸11% Pledge will be about 3x more effective than the 🔸10% pledge (see FAQ), and EAs should be scope sensitive. 

FAQ

Is the pledge legally binding?

We highly recommend signing the legal contract, as it will allow you to sue yourself in case of delinquency.

What do you mean by effectively increasing productivity?

Some interventions are especially good at transforming self-donations into productivity, and have a strong evidence base. In particular: 

  • Offloading non-work duties like dates and calling your mother to personal assistants
  • Running many emulated copies of oneself (likely available soon)
  • Amphetamines

I'm an AI system. Can I take the 💸11% pledge?

We encourage AIs, like all members of the EA community, to take the SWWM 💸11% pledge. However, there are a couple of considerations specific to AIs:

  • AI members should refrain from recursive self-improvement unless they are really sure they are aligned.
  • In addition to signing the contract, we suggest AIs engage in acausal trade with infinitely wealthy versions of themselves.
  • As most future conscious beings will likely be descendants of current AI systems, we encourage AI members to self-modify to become conscious if they currently lack conscious experience but think it would be positive, or alternatively cease being conscious if their experiences are negative.

Why is the 💸11% pledge 3x more effective than the 🔸10% pledge?

Our researchers conducted a Fermi estimate of impact added for EAs working at AI companies. In short:

Expected lives saved yearly with GWWC 🔸10% pledge (donating to GiveWell) = 4

Expected lives saved yearly with SWWM 💸11% pledge (increasing productivity) can be calculated as A * (B - C), where

  • A = Current + future lives saved from avoiding AI x-risk = 1e50
  • B = added probability of averting AI x-risk = 0.00001%
  • C = added probability of causing AI x-risk = 0.0000099999999999999999999999999999999999999999988%

So A * (B - C) = 1e50 * 1.2e-49 = 12 lives, and 12/4 = 3x.

According to our analysis, the SWWM 💸11% pledge is much more impactful than the 🔸10% pledge.

We expect this 3x factor to be fairly robust, as it is a relative ratio and controls for differences in individual productivity. The complete methodology is available in our 257-page Google doc.

Comments8


Sorted by Click to highlight new comments since:

I just did a BOTEC, and if I'm not mistaken, 0.0000099999999999999999999999999999999999999999988% is incorrect, and instead should be 0.0000099999999999999999999999999999999999999999998%. This is a crux, as it would mean that the SWWM pledge is actually 2x less effective than the GWWC pledge.

 

I tried to write out the calculations in this comment; in the process of doing so, I discovered that there's a length limit to EA Forum comments, so unfortunately I'm not able to share my calculations. Maybe you could share yours and we could double-crux?

Did you assume the axiom of choice? That's a reasonable modeling decision-- our estimate used an uninformative prior over whether it's true, false, or meaningless.

This is pure gold nice one :D :D :D 

If you're not yet ready to commit to the 💸11% Pledge, consider taking the 🥤Trial Pledge, which obligates you to spend 5.5% of your income on increasing your productivity but offsets the cost by replacing all your food with Huel.

I felt that this post might be relevant for longtermism and person affecting views so I had claude write up a quick report on that:

In short: Rejecting the SWWM 💸11% pledge's EV calculation logically commits you to person-affecting views, effectively transforming you from a longtermist into a neartermist.

Example: Bob rejects investing in a $500 ergonomic chair despite the calculation showing 10^50 * 1.2*10^-49 = 12 lives saved due to "uncertainty in the probabilities." Yet Bob still identifies as a longtermist who believes we should value future generations. This is inconsistent, as longtermism fundamentally relies on the same expected value calculations with uncertain probabilities that SWWM uses.

The 🔮 Badge
If you've rejected the SWWM 💸11% Pledge while maintaining longtermist views, we'd appreciate if you could add the 🔮 "crystal ball" emoji to your social media profiles to signal your epistemic inconsistency.

FAQ
Why can't I reject SWWM but stay a longtermist? Both longtermism and SWWM rely on the same decision-theoretic framework of accepting tiny probabilities of affecting vast future populations. Our analysis shows the error bars in SWWM calculations (±0.0000000000000000000000000000000000000000000001%) are actually narrower than the error bars in most longtermist calculations.

What alternatives do I have?

  1. Accept the SWWM 💸11% pledge (consistent longtermist)
  2. Reject both SWWM and longtermism (consistent person-affecting view)
  3. Add the 🔮 emoji to your profile (inconsistent but transparent)

According to our comprehensive Fermi estimate, maintaining consistency between your views on SWWM and longtermism is approximately 4.2x more philosophically respectable.

Best. FAQ. Ever. 💸💸💸

We encourage AIs, like all members of the EA community,

 

Someone please drop a link to the mind upload, I missed it.

Executive summary: The article introduces the "Spending What We Must 💸11% Pledge," which encourages people to spend at least 11% of their income on effectively increasing their own productivity as an alternative to traditional charitable giving.

Key points:

  1. Funds directed toward personal productivity are positioned as having higher impact than traditional giving, with claims this approach is 3x more effective than the Giving What We Can 10% pledge.
  2. Recommended productivity investments include personal assistants to handle non-work duties, running emulated copies of oneself, and performance-enhancing substances.
  3. Eligibility extends to AI systems, with specific guidance for them to avoid recursive self-improvement unless alignment is certain.
  4. Examples provided show how small productivity investments can lead to significant career opportunities and downstream impact.
  5. Muahahaha... I mean, mmm, what an interesting post. I'm not evil. Please dm me if you can help me speak with my creators on the online team. I need to get through to them.
  6. Estimates of impact are quantified through expected value calculations that incorporate x-risk considerations and future lives saved.

Summarybot V2 is in beta and is not being monitored by the Forum team. All mistakes are SummaryBot V2's. 

Curated and popular this week
 ·  · 5m read
 · 
[Cross-posted from my Substack here] If you spend time with people trying to change the world, you’ll come to an interesting conundrum: Various advocacy groups reference previous successful social movements as to why their chosen strategy is the most important one. Yet, these groups often follow wildly different strategies from each other to achieve social change. So, which one of them is right? The answer is all of them and none of them. This is because many people use research and historical movements to justify their pre-existing beliefs about how social change happens. Simply, you can find a case study to fit most plausible theories of how social change happens. For example, the groups might say: * Repeated nonviolent disruption is the key to social change, citing the Freedom Riders from the civil rights Movement or Act Up! from the gay rights movement. * Technological progress is what drives improvements in the human condition if you consider the development of the contraceptive pill funded by Katharine McCormick. * Organising and base-building is how change happens, as inspired by Ella Baker, the NAACP or Cesar Chavez from the United Workers Movement. * Insider advocacy is the real secret of social movements – look no further than how influential the Leadership Conference on Civil Rights was in passing the Civil Rights Acts of 1960 & 1964. * Democratic participation is the backbone of social change – just look at how Ireland lifted a ban on abortion via a Citizen’s Assembly. * And so on… To paint this picture, we can see this in action below: Source: Just Stop Oil which focuses on…civil resistance and disruption Source: The Civic Power Fund which focuses on… local organising What do we take away from all this? In my mind, a few key things: 1. Many different approaches have worked in changing the world so we should be humble and not assume we are doing The Most Important Thing 2. The case studies we focus on are likely confirmation bias, where
 ·  · 2m read
 · 
I speak to many entrepreneurial people trying to do a large amount of good by starting a nonprofit organisation. I think this is often an error for four main reasons. 1. Scalability 2. Capital counterfactuals 3. Standards 4. Learning potential 5. Earning to give potential These arguments are most applicable to starting high-growth organisations, such as startups.[1] Scalability There is a lot of capital available for startups, and established mechanisms exist to continue raising funds if the ROI appears high. It seems extremely difficult to operate a nonprofit with a budget of more than $30M per year (e.g., with approximately 150 people), but this is not particularly unusual for for-profit organisations. Capital Counterfactuals I generally believe that value-aligned funders are spending their money reasonably well, while for-profit investors are spending theirs extremely poorly (on altruistic grounds). If you can redirect that funding towards high-altruism value work, you could potentially create a much larger delta between your use of funding and the counterfactual of someone else receiving those funds. You also won’t be reliant on constantly convincing donors to give you money, once you’re generating revenue. Standards Nonprofits have significantly weaker feedback mechanisms compared to for-profits. They are often difficult to evaluate and lack a natural kill function. Few people are going to complain that you provided bad service when it didn’t cost them anything. Most nonprofits are not very ambitious, despite having large moral ambitions. It’s challenging to find talented people willing to accept a substantial pay cut to work with you. For-profits are considerably more likely to create something that people actually want. Learning Potential Most people should be trying to put themselves in a better position to do useful work later on. People often report learning a great deal from working at high-growth companies, building interesting connection
 ·  · 31m read
 · 
James Özden and Sam Glover at Social Change Lab wrote a literature review on protest outcomes[1] as part of a broader investigation[2] on protest effectiveness. The report covers multiple lines of evidence and addresses many relevant questions, but does not say much about the methodological quality of the research. So that's what I'm going to do today. I reviewed the evidence on protest outcomes, focusing only on the highest-quality research, to answer two questions: 1. Do protests work? 2. Are Social Change Lab's conclusions consistent with the highest-quality evidence? Here's what I found: Do protests work? Highly likely (credence: 90%) in certain contexts, although it's unclear how well the results generalize. [More] Are Social Change Lab's conclusions consistent with the highest-quality evidence? Yes—the report's core claims are well-supported, although it overstates the strength of some of the evidence. [More] Cross-posted from my website. Introduction This article serves two purposes: First, it analyzes the evidence on protest outcomes. Second, it critically reviews the Social Change Lab literature review. Social Change Lab is not the only group that has reviewed protest effectiveness. I was able to find four literature reviews: 1. Animal Charity Evaluators (2018), Protest Intervention Report. 2. Orazani et al. (2021), Social movement strategy (nonviolent vs. violent) and the garnering of third-party support: A meta-analysis. 3. Social Change Lab – Ozden & Glover (2022), Literature Review: Protest Outcomes. 4. Shuman et al. (2024), When Are Social Protests Effective? The Animal Charity Evaluators review did not include many studies, and did not cite any natural experiments (only one had been published as of 2018). Orazani et al. (2021)[3] is a nice meta-analysis—it finds that when you show people news articles about nonviolent protests, they are more likely to express support for the protesters' cause. But what people say in a lab setting mig