Hide table of contents

Say I have raised $20k-$400k for a new EA project/organization. I want to spend that money on talent and services. What kind of legal entity should I use for this? Is there any good advice on how to go about setting one up for this purpose, and how to maintain it after it's set up?

This applies to me, but I imagine that there may be many other new groups considering or planning setting up the legal and financial infrastructure for relatively-small projects and organizations.

Some options:

A US 501(c)(3)

A 501(c)(3) would allow for tax deductibility from American donors. However, from what I understand, these are the most expensive and difficult organizations to set up and manage. At some point in an organization's life, it probably makes sense to register, but this seems like a lot of work during the beginning, especially if it is not obvious how long the project will last or what will become.

Get sponsored by an existing 501(c)(3)

Getting sponsored by an existing 501(c)(3) would allow for tax-deductible donations and may help with some of the operations. However, is not obvious who in the EA community would do this. I have not heard of many EA organizations sponsoring other groups. Sponsorship would be a liability to the sponsoring organization, so I imagine groups would be quite hesitant.

An LLC

Right now an LLC seems like generally the best option to me, but I'm not sure. It doesn't get tax deductibility, but this may not be a big deal if much of the money is coming from nonprofits (as would be true for BERI or Effective Altruist Funds grants). The Open Philanthropy Project set up as an LLC for the flexibility it provides, which I find interesting.

I imagine that setting up a Delaware LLC would be reasonable. Stripe Atlas seems like an interesting option for setting one up, though is geared more for internet companies and is a bit expensive.

Sole Proprietorship

This is probably the default option for small amounts, but definitely has some downsides. For one, it doesn't limit liability, and for another, the accounting could be quite messy.

14

0
0

Reactions

0
0
Comments3


Sorted by Click to highlight new comments since:

Are you considering European structures as well, or is it limited to US?

I'm personally more interested in US structures, mainly because I have a lot more familiarity with them and expect to spend most of my career in the US. That said, this post was meant to collect general advice for others also doing similar things, so any thoughts are appreciated.

There are charities that specialize in sponsoring projects to give them tax deductability, one being the Social and Environmental Entrepreneurs.

More from Ozzie Gooen
82
Ozzie Gooen
· · 9m read
Curated and popular this week
 ·  · 52m read
 · 
In recent months, the CEOs of leading AI companies have grown increasingly confident about rapid progress: * OpenAI's Sam Altman: Shifted from saying in November "the rate of progress continues" to declaring in January "we are now confident we know how to build AGI" * Anthropic's Dario Amodei: Stated in January "I'm more confident than I've ever been that we're close to powerful capabilities... in the next 2-3 years" * Google DeepMind's Demis Hassabis: Changed from "as soon as 10 years" in autumn to "probably three to five years away" by January. What explains the shift? Is it just hype? Or could we really have Artificial General Intelligence (AGI) by 2028?[1] In this article, I look at what's driven recent progress, estimate how far those drivers can continue, and explain why they're likely to continue for at least four more years. In particular, while in 2024 progress in LLM chatbots seemed to slow, a new approach started to work: teaching the models to reason using reinforcement learning. In just a year, this let them surpass human PhDs at answering difficult scientific reasoning questions, and achieve expert-level performance on one-hour coding tasks. We don't know how capable AGI will become, but extrapolating the recent rate of progress suggests that, by 2028, we could reach AI models with beyond-human reasoning abilities, expert-level knowledge in every domain, and that can autonomously complete multi-week projects, and progress would likely continue from there.  On this set of software engineering & computer use tasks, in 2020 AI was only able to do tasks that would typically take a human expert a couple of seconds. By 2024, that had risen to almost an hour. If the trend continues, by 2028 it'll reach several weeks.  No longer mere chatbots, these 'agent' models might soon satisfy many people's definitions of AGI — roughly, AI systems that match human performance at most knowledge work (see definition in footnote).[1] This means that, while the co
saulius
 ·  · 22m read
 · 
Summary In this article, I estimate the cost-effectiveness of five Anima International programs in Poland: improving cage-free and broiler welfare, blocking new factory farms, banning fur farming, and encouraging retailers to sell more plant-based protein. I estimate that together, these programs help roughly 136 animals—or 32 years of farmed animal life—per dollar spent. Animal years affected per dollar spent was within an order of magnitude for all five evaluated interventions. I also tried to estimate how much suffering each program alleviates. Using SADs (Suffering-Adjusted Days)—a metric developed by Ambitious Impact (AIM) that accounts for species differences and pain intensity—Anima’s programs appear highly cost-effective, even compared to charities recommended by Animal Charity Evaluators. However, I also ran a small informal survey to understand how people intuitively weigh different categories of pain defined by the Welfare Footprint Institute. The results suggested that SADs may heavily underweight brief but intense suffering. Based on those findings, I created my own metric DCDE (Disabling Chicken Day Equivalent) with different weightings. Under this approach, interventions focused on humane slaughter look more promising, while cage-free campaigns appear less impactful. These results are highly uncertain but show how sensitive conclusions are to how we value different kinds of suffering. My estimates are highly speculative, often relying on subjective judgments from Anima International staff regarding factors such as the likelihood of success for various interventions. This introduces potential bias. Another major source of uncertainty is how long the effects of reforms will last if achieved. To address this, I developed a methodology to estimate impact duration for chicken welfare campaigns. However, I’m essentially guessing when it comes to how long the impact of farm-blocking or fur bans might last—there’s just too much uncertainty. Background In
 ·  · 4m read
 · 
SUMMARY:  ALLFED is launching an emergency appeal on the EA Forum due to a serious funding shortfall. Without new support, ALLFED will be forced to cut half our budget in the coming months, drastically reducing our capacity to help build global food system resilience for catastrophic scenarios like nuclear winter, a severe pandemic, or infrastructure breakdown. ALLFED is seeking $800,000 over the course of 2025 to sustain its team, continue policy-relevant research, and move forward with pilot projects that could save lives in a catastrophe. As funding priorities shift toward AI safety, we believe resilient food solutions remain a highly cost-effective way to protect the future. If you’re able to support or share this appeal, please visit allfed.info/donate. Donate to ALLFED FULL ARTICLE: I (David Denkenberger) am writing alongside two of my team-mates, as ALLFED’s co-founder, to ask for your support. This is the first time in Alliance to Feed the Earth in Disaster’s (ALLFED’s) 8 year existence that we have reached out on the EA Forum with a direct funding appeal outside of Marginal Funding Week/our annual updates. I am doing so because ALLFED’s funding situation is serious, and because so much of ALLFED’s progress to date has been made possible through the support, feedback, and collaboration of the EA community.  Read our funding appeal At ALLFED, we are deeply grateful to all our supporters, including the Survival and Flourishing Fund, which has provided the majority of our funding for years. At the end of 2024, we learned we would be receiving far less support than expected due to a shift in SFF’s strategic priorities toward AI safety. Without additional funding, ALLFED will need to shrink. I believe the marginal cost effectiveness for improving the future and saving lives of resilience is competitive with AI Safety, even if timelines are short, because of potential AI-induced catastrophes. That is why we are asking people to donate to this emergency appeal