We've had a lot of votes on the banner! If you'd like to explain why you voted the way you did, what your cruxes[1] are, and what would change your mind, comment in this thread. 

You can also mention if you'd be open for having a dialogue with another Forum user who disagrees with you. If someone comments and offers to dialogue with you, you can set up a time to write a dialogue together (perhaps via Forum dms). 

To find out more about the event, and how to contribute, read the announcement post.

  1. ^

    Beliefs or assumptions which determine your overall opinion, but are better targets for argument/ you would more easily change your mind on. For example, one of mine is "philosophy of mind doesn't make progress". 

28

1
0

Reactions

1
0
Comments24


Sorted by Click to highlight new comments since:

Ok--at Toby's encouragement, here are my thoughts:

This is a very old point, but to my mind, at least from a utilitarian perspective, the main reason it's worth working on promoting AI welfare is the risk of foregone upside. I.e. without actively studying what constitutes AI welfare and advocating for producing it, we seem likely to have a future that's very comfortable for ourselves and our descendants--fully automated luxury space communism, if you like--but which contains a very small proportion of the value that could have been created by creating lots of happy artificial minds. So concern for creating AI welfare seems likely to be the most important way in which utilitarian and human-common-sense moral recommendations differ.

It seems to me that the amount of value we could create if we really optimized for total AI welfare is probably greater than the amount of disvalue we'll create if we just use AI tools and allow for suffering machines by accident, since in the latter case the suffering would be a byproduct, not something anyone optimizes for.

But AI welfare work (especially if this includes moral advocacy) just for the sake of avoiding this downside also seems valuable enough to be worth a lot of effort on its own, even if suffering AI tools are a long way off. The animal analogy seems relevant: it's hard to replace factory farming once people have started eating a lot of meat, but in India, where Hinduism has discouraged meat consumption for a long time, less meat is consumed and so factory farming is evidently less widespread.

So in combination, I expect AI welfare work of some kind or another is probably very important. I have almost no idea what the best interventions would be or how cost-effective they would be, so I have no opinion on exactly how much work should go into them. I expect no one really knows at this point. But at face value the topic seems important enough to warrant at least doing exploratory work until we have a better sense of what can be done and how cost-effective it could be, only stopping in the (I think unlikely) event that we can say with some confidence that the best AI welfare work to be done is worse than the best work that can be done in other areas.

When telling stories like your first paragraph, I wish people either said "almost all of the galaxies we reach are tiled with some flavor of computronium and here's how AI welfare work affected the flavor" or "it is not the case that almost all of the galaxies we reach are tiled with some flavor of computronium and here's why."

The universe will very likely be tiled with some flavor of computronium is a crucial consideration, I think.

To my mind, the first point applies to whatever resources are used throughout the future, whether it’s just the earth or some larger part of the universe.

I agree that the number/importance of welfare subjects in the future is a crucial consideration for how much to do longtermist as opposed to other work. But when comparing longtermist interventions—say, splitting a budget between lowering the risk of the world ending and proportionally increasing the fraction of resources devoted to creating happy artificial minds—it would seem to me that the “size of the future” typically multiplies the value of both interventions equally, and so doesn’t matter.

kta
15
3
0

(Not an AI welfare/safety expert by any stretch, just adding my two cents here! Also I was very piqued by the banner and loved hovering over the footnote! I've thought about digital sentience, but this banner and this week really put me into a "hmm..." state)

My view leans towards "moderately disagree." (I fluctuated between this, neutral, and slightly agree.) For context, when it's AI safety, I'd say "highly agree." Thoughts behind my current position:

Why I'd prioritize it less: 

  • I consider myself longtermist, but I have always grappled with the opportunity costs of highly prioritizing more "speculative" areas. I care about high EV areas, I also grapple with deprioritizing very tangible cause areas with existing beings that have high EV too. When I looked at the table below, I'd lean towards giving more resources towards AW versus making AI welfare right now a priority. 
  • I also think about how, if we divert more resources into AI welfare, I worry about the ramifications of diverting more EAs into a very dense, specialized sector. While this specialization is important, I am concerned that it might sometimes lead to a narrower focus that doesn't fully account for the broader, interconnected systems of the world. In contrast, fields like biosecurity often consider a wider range of factors and have a more integrative perspective. This more holistic view can be crucial in addressing complex, multifaceted issues, and one reason I would prioritize AI welfare less is the opportunity costs towards areas that may be more holistic (not saying AI welfare has no reason to be considered holistic)
  • I have some concerns that trying to help AI right now might make things worse since we don't fully know yet what's being done now that can make things riskier? (Nathan said something to this effect in this thread).
  • I don't know to what extent AI welfare is irreversible compared to unaligned AI
  • It seems less likely for multiplanetary civilizations to develop with advanced AI, reducing likelihood of AI systems across the universe, which reduces my prioritizing AI welfare on a universal scale

Why I'd still prioritize it: 

  • I can't see myself prescribing a 0% chance AI would be sentient, and I can't see myself prescribing less than (edit:) 2% of resources and talent in effective altruism to something wide-scale I'd hold a possibility of being considered sentient, even if it might be less standard (i.e. more outside average moral circles) because of big value creation, just generally preventing suffering, and potentially preventing additional happiness, all of which I'm highly for.
  • I think exploratory and not very tapped-in work needs to be done more, and just establishing enough baseline infrastructure is important for this high EV type of cause (assuming we would say AI will be very widespread)
  • I like the trammell's animal welfare analogy

Overall, I agree that resources and talent should be allocated to AI welfare because it's prudent and can prevent future suffering. However, I moderately disagree with it being an EA priority due to its current speculative nature and how I think AI safety. I think AI safety and solving the alignment problem should be a priority, especially in these next few years, though, and hold some confidence in preventing digital suffering. 

Other thoughts:

  • I wonder if there'd ever be a conflict between AI welfare and human welfare or the welfare of other beings. Haven't put much thought here. Something that immediately comes to mind is if advanced AI systems would require substantial energy and infrastructure, potentially competing with human needs. From a utilitarian point of view, this presents a significant dilemma. However, there's the argument that solving AI alignment could mitigate these issues, ensuring that AI systems are developed and managed in ways that do not harm human welfare. My current thinking is that there's less likely potential for conflict between AI and human welfare if we solve the alignment problem and improve the policy infrastructure around AI. I might compare bioethics to historical precedents, showing that ethical alignment leads to better welfare outcomes
  • Some media that have made me truly feel for AI welfare are "I, Robot," "Her," Black Mirror's "Joan is Awful," and "Klara and the Sun”!

My reasons for not making it a priority:

  • it seems like we will be better placed to solve the issue in the future (we will understand AIs much better specifically, and also perhaps will just have much better intellectual tools generally),
  • it seems like most of the mistakes we can make by getting this wrong are mistakes we can fix if we get it right later on.

By contrast to existential risk, which we need to get right now or lose the opportunity (and all other opportunities) forever, I don't see a corresponding loss of option value here. Perhaps it's worth thinking about how to ensure we preserve the will to solve the issue through whatever upheaval comes next. But I think that's much easier than actually trying to solve it right now.

edit: I think the first consideration isn't nearly as strong for poverty / health interventions and animal interventions: it feels more like we already know some good things to do there so I'm on board with starting now, especially in cases where we think their effects will compound over time.

Do you have a sense of what you think the right amout to spend is?

I think spending zero dollars (and hours) isn't obviously a mistake, but I'd be willing to listen to someone who wanted to advocate for some specific intervention to be funded.

It seems really valuable to have experts at the time the discussion happens. 

If you agree, then it seems worth trianing people for the future when we discuss it.

We can do that in the future too?

Training takes probably 3 years to cycle up and maybe 3 years to happen. When did we start deciding to train people in AI Safety, vs when was there enough?
 

Seems plausible to me that the AI welfare discussion happens before we might currently be ready.

But again you're suggesting a time-limited window in which the AI welfare discussion happens, and if we don't intervene in that window it'll be too late. I just don't see a reason to expect that. I imagine that after the first AI welfare discussion, there will be others. While obviously it's best to fix problems as soon as possible, I don't think it's worth diverting resources from problems with a much clearer reason to believe we need to act now.

I think that being there at the start of a discussion is a great way to shift it. Look at AI safety (for good and ill)

For me, a key question is "How much is 5%?". 

Here is a table I found.

So it seems like right now 5% is somewhere in the same range as Animal Welfare and EA Meta funding.

I guess that seems a bit high, given that animals exist and AIs don't.

I think a key benefit of AI work was training AI Safety folks to be around when needed. Having resources at a crucial moment isn't solely about money, it's about having the resource that is useful in that moment. A similar thing to do might be to train philosophers and government staffers and activists who are well versed in the AI welfare arguments who can act if need be. 

Not clear to me that that requires 5% of EA funding though. 

this is super helpful! would be cool if we can see %s given to insect sentience or other smaller sub cause areas like that. does anyone have access to that?

I'd guess a less than .5% (90%)

I think the burden of proof lies with those advocating for AI welfare as an EA priority.

So far, I haven't read compelling arguments to change my default.

I agree and I hope we get some strong arguments from those in favor. I would imagine there is already a bunch of stuff written given the recent Open Phil defunding it kerfuffle.

What's your thought on this:

Humans are conscious in whatever way that word is normally used

  • Our brains are made of matter
  • I think it's likely we'll be able to use matter to make other conscious minds
  • These minds may be able to replicate far faster than our own
  • A huge amount of future consciousness may be non-human
  • The wellbeing of huge chunks of funture consciousness are worthy of our concern

I think it's likely we'll be able to use matter to make other conscious minds

Can you expand on this? Do you think that a model loaded onto a GPU could be conscious?

And do you think bacteria might be conscious?

I think given a big enough GPU, yes, it seems plausible to me. Our mids are memory stores and performing calculations. What is missing in terms of a GPU?

I think bacteria are unlikely to be conscious due to a lack of processing power. 

What is missing in terms of a GPU?

Something unknown.

I think given a big enough GPU, yes, it seems plausible to me. Our mids are memory stores and performing calculations.

Do you think it's plausible that a GPU rendering graphics is conscious? Or do you think that a GPU can only be conscious when it runs a model that mimics human behavior?

I think bacteria are unlikely to be conscious due to a lack of processing power.

Potential counterargument: microbial intelligence.

I have doubts about whether digital or electronic information flows will yield valenced sentience, although I don't rule it out.

But I have much stronger doubts about whether we can ever know what makes these 'digital minds' happy or sad, or even what goes in what direction.

Despite being a panpsychist, I rate it fairly low. I don't see a future where we solve AI safety where there are a lot of suffering AIs. If we fail on safety, then it wont matter what you wrote about AI welfare, the unaligned AI is not going to be moved by it.

My thoughts on the issue:

  • Humans are conscious in whatever way that word is normally used
  • Our brains are made of matter
  • I think it's likely we'll be able to use matter to make other conscious minds
  • These minds may be able to replicate far faster than our own
  • A huge amount of future consciousness may be non-human
  • The wellbeing of huge chunks of funture consciousness are worthy of our concern
  • Forecasting is hard and it's not clear how valuable early AI safety work was (maybe it led to much more risk ??)
  • Early work on AI welfare might quite plausibly make the problem worse. 
  • Part of me would like to get better calibrated on which interventions work and which don't before hugely focusing on such an important issue
  • Part of me would like to fund general state capacity work and train experts to be in a better place when this happens
Curated and popular this week
 ·  · 20m read
 · 
Advanced AI could unlock an era of enlightened and competent government action. But without smart, active investment, we’ll squander that opportunity and barrel blindly into danger. Executive summary See also a summary on Twitter / X. The US federal government is falling behind the private sector on AI adoption. As AI improves, a growing gap would leave the government unable to effectively respond to AI-driven existential challenges and threaten the legitimacy of its democratic institutions. A dual imperative → Government adoption of AI can’t wait. Making steady progress is critical to: * Boost the government’s capacity to effectively respond to AI-driven existential challenges * Help democratic oversight keep up with the technological power of other groups * Defuse the risk of rushed AI adoption in a crisis → But hasty AI adoption could backfire. Without care, integration of AI could: * Be exploited, subverting independent government action * Lead to unsafe deployment of AI systems * Accelerate arms races or compress safety research timelines Summary of the recommendations 1. Work with the US federal government to help it effectively adopt AI Simplistic “pro-security” or “pro-speed” attitudes miss the point. Both are important — and many interventions would help with both. We should: * Invest in win-win measures that both facilitate adoption and reduce the risks involved, e.g.: * Build technical expertise within government (invest in AI and technical talent, ensure NIST is well resourced) * Streamline procurement processes for AI products and related tech (like cloud services) * Modernize the government’s digital infrastructure and data management practices * Prioritize high-leverage interventions that have strong adoption-boosting benefits with minor security costs or vice versa, e.g.: * On the security side: investing in cyber security, pre-deployment testing of AI in high-stakes areas, and advancing research on mitigating the ris
 ·  · 15m read
 · 
In our recent strategy retreat, the GWWC Leadership Team recognised that by spreading our limited resources across too many projects, we are unable to deliver the level of excellence and impact that our mission demands. True to our value of being mission accountable, we've therefore made the difficult but necessary decision to discontinue a total of 10 initiatives. By focusing our energy on fewer, more strategically aligned initiatives, we think we’ll be more likely to ultimately achieve our Big Hairy Audacious Goal of 1 million pledgers donating $3B USD to high-impact charities annually. (See our 2025 strategy.) We’d like to be transparent about the choices we made, both to hold ourselves accountable and so other organisations can take the gaps we leave into account when planning their work. As such, this post aims to: * Inform the broader EA community about changes to projects & highlight opportunities to carry these projects forward * Provide timelines for project transitions * Explain our rationale for discontinuing certain initiatives What’s changing  We've identified 10 initiatives[1] to wind down or transition. These are: * GWWC Canada * Effective Altruism Australia funding partnership * GWWC Groups * Giving Games * Charity Elections * Effective Giving Meta evaluation and grantmaking * The Donor Lottery * Translations * Hosted Funds * New licensing of the GWWC brand  Each of these is detailed in the sections below, with timelines and transition plans where applicable. How this is relevant to you  We still believe in the impact potential of many of these projects. Our decision doesn’t necessarily reflect their lack of value, but rather our need to focus at this juncture of GWWC's development.  Thus, we are actively looking for organisations and individuals interested in taking on some of these projects. If that’s you, please do reach out: see each project's section for specific contact details. Thank you for your continued support as we
 ·  · 3m read
 · 
We are excited to share a summary of our 2025 strategy, which builds on our work in 2024 and provides a vision through 2027 and beyond! Background Giving What We Can (GWWC) is working towards a world without preventable suffering or existential risk, where everyone is able to flourish. We do this by making giving effectively and significantly a cultural norm. Focus on pledges Based on our last impact evaluation[1], we have made our pledges –  and in particular the 🔸10% Pledge – the core focus of GWWC’s work.[2] We know the 🔸10% Pledge is a powerful institution, as we’ve seen almost 10,000 people take it and give nearly $50M USD to high-impact charities annually. We believe it could become a norm among at least the richest 1% — and likely a much wider segment of the population — which would cumulatively direct an enormous quantity of financial resources towards tackling the world’s most pressing problems.  We initiated this focus on pledges in early 2024, and are doubling down on it in 2025. In line with this, we are retiring various other initiatives we were previously running and which are not consistent with our new strategy. Introducing our BHAG We are setting ourselves a long-term Big Hairy Audacious Goal (BHAG) of 1 million pledgers donating $3B USD to high-impact charities annually, which we will start working towards in 2025. 1 million pledgers donating $3B USD to high-impact charities annually would be roughly equivalent to ~100x GWWC’s current scale, and could be achieved by 1% of the world’s richest 1% pledging and giving effectively. Achieving this would imply the equivalent of nearly 1 million lives being saved[3] every year. See the BHAG FAQ for more info. Working towards our BHAG Over the coming years, we expect to test various growth pathways and interventions that could get us to our BHAG, including digital marketing, partnerships with aligned organisations, community advocacy, media/PR, and direct outreach to potential pledgers. We thin