Hide table of contents

The FTX Foundation's Future Fund is a philanthropic fund making grants and investments to ambitious projects in order to improve humanity's long-term prospects.

Our regranting program will offer discretionary budgets to independent part-time grantmakers, to be spent in the next ~6 months. Budgets will typically be in the $250k-few million range. We've already invited a first cohort of 21 regrantors to test the program.

Large funders aren’t always aware of the best opportunities. We want to try a decentralized approach. We hope to empower a range of interesting, ambitious, and altruistic people to drive funding decisions through a rewarding, low-friction process. 

We are just getting started and we'd like to fund a lot of great projects. So our primary goal for 2022 is to perform bold tests of new approaches to scaling grantmaking. We very much consider this an experiment, and our goal is to decisively test whether this kind of program works. 

If successful, this program will help us identify great grants that we would have missed, help new people launch exciting projects, and find and empower people who could be strong grantmakers.

We're also aware that the program could cause various problems. We'll try to minimize these downsides by being thoughtful about who we select, providing relevant guidance, and carefully screening grant recommendations for downside risks and conflicts of interest.

We’re so excited to see what regrantors come up with!

Become a regrantor 

We’re planning to invite additional regrantors to join the program in a month or so. Some of these will come from our existing networks, but we’re also opening up a public process to be considered as a regrantor. 

If you’re interested, please fill these out as soon as possible; we’re reviewing materials on a rolling basis. 

About the role

Regrantors will be assigned a discretionary budget, from which they can make grant recommendations. These will be screened primarily for downside risk, conflicts of interest, and consistency with our charitable mission and tax-exempt status. An exception is that we are less likely to approve funding for orgs that are already standard targets for EA funders, because we mainly want to fund new things.

The regranting pot expires after 6 months. Regrantors will be compensated for their work based on the quality and volume of their grantmaking.

The role is extremely independent. While we'll provide some documents with basic guidance, regrantors are expected to come up with all of their own ideas for the grants they recommend, do all the necessary communication with grantees, and ensure the quality of their own work. Once the grant recommendation is reviewed and approved, we'll take care of the grantee getting the money—but that's all we're planning to provide for this program.

While we will fund their grants, regrantors will not be Future Fund/FTX Foundation employees or be authorized to speak for the Future Fund/FTX Foundation.

What we’re looking for

We're looking for people who have great grant ideas that we're likely to miss. 

We’ll also lean towards giving this opportunity to people who don’t already have easy access to funds, and those for whom it might make a valuable learning experience. (The regrantors we invite won’t just be a “who’s who” of EA—we expect that many of the effective altruists we most admire will receive no regranting pot at all.)

We think excellent regrantors will:

  • Have expertise related to our areas of interest and project ideas.
  • Have good judgment about people and projects.
  • Leverage their diverse networks.
  • Be proactive and have an entrepreneurial drive. We think the best grants will require actively going out and helping things get started, or finding new people.
  • Experiment with new things, to teach us about new approaches to grantmaking or areas we aren't familiar with.

In addition, we ask that regrantors:

  • Communicate professionally.
  • Create a good experience for grantees and potential grantees.
  • Be sensitive to the effects their funding has on the ecosystems they work in.

Types of grants we’re especially interested in:

  • Incubating new projects. We’d love to see regrantors help seed new projects. We’re especially interested in new projects that could eventually scale up dramatically. (Read more about our vision for massively scalable projects here.)
  • Bringing in new people. We’re keen on regrantors finding talented people, who they believe have promise for making an effective contribution, and helping them get started.
  • New approaches. We’re interested in getting our feet wet on new types of projects even if they haven’t received much longtermist attention so far.
  • Local knowledge. Are there easy opportunities to support the work of people in the regrantor’s network? Is there a specific, tractable subproblem in the regrantor’s field, for which they could announce a big prize?

We’re happy to answer questions, though it might take us a few days to respond due to other programs and content we're launching right now.

We look forward to hearing from you!



 



 







 

Comments11


Sorted by Click to highlight new comments since:

Cool program! One question on this:

We've already invited a first cohort of 21 regrantors to test the program.

Will you be announcing who these regrantors are, and when will that happen if so?

Thanks! We are not planning to publish the list of regrantors for now.

I'd be very interested  in joining as a regranter, though it may make sense to wait a few years, by which point I will have donated most of my crypto pool and gained a bunch of experience. You can see my current strategy at Being an individual alignment grantmaker.

Edit: Does screening for conflicts of interest mean not allowing regranters to grant to people they know? If yes, I can see the reasoning, but if I was operating under this rule it would have blocked several of my most promising grants, which I found through personal connections. I would propose having these grants marked clearly and the regranter's reputation being more strongly staked on those grants going well, rather than outright banning them.

Edit2: Will there be a network for regranters (e.g. Discord, Slack), and would it be possible for me to join as an independent grantmaker to share knowledge and best practices? Or maybe I should just apply now as I'm keen to learn, just not confident I am ready to direct $250k+/year.

You are welcome to apply now!

Regrantors are able to make grants to people they know (in fact, having a diverse network is part of what makes for an effective regrantor); they just have to disclose if there's a conflict of interest, and we may reject a grant if we don't feel comfortable with it on those grounds. 

We don't currently have a network for regrantors that is open for external people to join.

This is very exciting!

For those interested in applying to to become a regrantor, is there a deadline? And even if there's no hard deadline, is there a time that would be useful to apply by?

We’re planning to invite additional regrantors by the end of this month or so. We are evaluating regrantor expressions of interest/referrals for regrantors on a rolling basis, so please send these in as soon as possible.

Good to know, thanks!

Is there compensation for the regrantors?

Yes: "Regrantors will be compensated for their work based on the quality and volume of their grantmaking."

Is this program completely done / finished now?

As in, are regrantors no longer able to make grants?

Is it appropriate for people seeking funding to seek out regrantors in order to submit proposals to them? 

Given that you aren't willing to publish the list of regrantors, this makes me suspect the answer is "no".

Curated and popular this week
Ben_West🔸
 ·  · 1m read
 · 
> Summary: We propose measuring AI performance in terms of the length of tasks AI agents can complete. We show that this metric has been consistently exponentially increasing over the past 6 years, with a doubling time of around 7 months. Extrapolating this trend predicts that, in under a decade, we will see AI agents that can independently complete a large fraction of software tasks that currently take humans days or weeks. > > The length of tasks (measured by how long they take human professionals) that generalist frontier model agents can complete autonomously with 50% reliability has been doubling approximately every 7 months for the last 6 years. The shaded region represents 95% CI calculated by hierarchical bootstrap over task families, tasks, and task attempts. > > Full paper | Github repo Blogpost; tweet thread. 
 ·  · 2m read
 · 
For immediate release: April 1, 2025 OXFORD, UK — The Centre for Effective Altruism (CEA) announced today that it will no longer identify as an "Effective Altruism" organization.  "After careful consideration, we've determined that the most effective way to have a positive impact is to deny any association with Effective Altruism," said a CEA spokesperson. "Our mission remains unchanged: to use reason and evidence to do the most good. Which coincidentally was the definition of EA." The announcement mirrors a pattern of other organizations that have grown with EA support and frameworks and eventually distanced themselves from EA. CEA's statement clarified that it will continue to use the same methodologies, maintain the same team, and pursue identical goals. "We've found that not being associated with the movement we have spent years building gives us more flexibility to do exactly what we were already doing, just with better PR," the spokesperson explained. "It's like keeping all the benefits of a community while refusing to contribute to its future development or taking responsibility for its challenges. Win-win!" In a related announcement, CEA revealed plans to rename its annual EA Global conference to "Coincidental Gathering of Like-Minded Individuals Who Mysteriously All Know Each Other But Definitely Aren't Part of Any Specific Movement Conference 2025." When asked about concerns that this trend might be pulling up the ladder for future projects that also might benefit from the infrastructure of the effective altruist community, the spokesperson adjusted their "I Heart Consequentialism" tie and replied, "Future projects? I'm sorry, but focusing on long-term movement building would be very EA of us, and as we've clearly established, we're not that anymore." Industry analysts predict that by 2026, the only entities still identifying as "EA" will be three post-rationalist bloggers, a Discord server full of undergraduate philosophy majors, and one person at
Thomas Kwa
 ·  · 2m read
 · 
Epistemic status: highly certain, or something The Spending What We Must 💸11% pledge  In short: Members pledge to spend at least 11% of their income on effectively increasing their own productivity. This pledge is likely higher-impact for most people than the Giving What We Can 🔸10% Pledge, and we also think the name accurately reflects the non-supererogatory moral beliefs of many in the EA community. Example Charlie is a software engineer for the Centre for Effective Future Research. Since Charlie has taken the SWWM 💸11% pledge, rather than splurge on a vacation, they decide to buy an expensive noise-canceling headset before their next EAG, allowing them to get slightly more sleep and have 104 one-on-one meetings instead of just 101. In one of the extra three meetings, they chat with Diana, who is starting an AI-for-worrying-about-AI company, and decide to become a cofounder. The company becomes wildly successful, and Charlie's equity share allows them to further increase their productivity to the point of diminishing marginal returns, then donate $50 billion to SWWM. The 💸💸💸 Badge If you've taken the SWWM 💸11% Pledge, we'd appreciate if you could add three 💸💸💸 "stacks of money with wings" emoji to your social media profiles. We chose three emoji because we think the 💸11% Pledge will be about 3x more effective than the 🔸10% pledge (see FAQ), and EAs should be scope sensitive.  FAQ Is the pledge legally binding? We highly recommend signing the legal contract, as it will allow you to sue yourself in case of delinquency. What do you mean by effectively increasing productivity? Some interventions are especially good at transforming self-donations into productivity, and have a strong evidence base. In particular:  * Offloading non-work duties like dates and calling your mother to personal assistants * Running many emulated copies of oneself (likely available soon) * Amphetamines I'm an AI system. Can I take the 💸11% pledge? We encourage A