TK

Thomas Kwa

Member of Technical Staff @ METR
3203 karmaJoined Working (0-5 years)Berkeley, CA, USA

Bio

Participation
4

AI safety researcher

Comments
259

The Pulse survey has now basically allayed all of my concerns.

Thanks, I've started donating $33/month to the FarmKind bonus fund, which is double the calculator estimate for my diet. [1] I will probably donate ~$10k of stocks in 2025 to offset my lifetime diet impact-- is there any reason not to do this? I've already looked at the non-counterfactual matching argument, which I don't find convincing.

[1] I basically never eat chicken, substituting it with other meats, so I reduced the poultry category by 2/3 and allocated that proportionally between the beef and pork categories.

I disagree with a few points, especially paragraph 1. Are you saying that people were worried about abolition slowing down economic growth and lowering standards of living? I haven't heard this as a significant concern-- free labor was perfectly capable of producing cotton at a small premium, and there were significant British boycotts of slave-produced products like cotton and sugar.

As for utilitarian arguments, that's not the main way I imagine EAs would help. EA pragmatists would prioritize the cause for utilitarian reasons and do whatever is best to achieve their policy goals, much as we are already doing for animal welfare. The success of EAs in animal welfare, or indeed anywhere other than x-risk, is in implementation of things like corporate campaigns rather than mass spreading of arguments. Even in x-risk, an alliance with natsec people has effected concrete policy outcomes like compute export controls.

To paragraph 2, the number of philosophers is pretty low in contemporary EA. We just hear about them more. And while abolition might have been relatively intractable in the US, my guess is the UK could have been sped up. 

I basically agree with paragraph 3, though I would hope if it came to it we would find something more economical than directly freeing slaves.

Overall thanks for the thoughtful response! I wouldn't mind discussing this more.

I was imagining a split similar to the present, in which over half of EAs were American or British.

How do I offset my animal product consumption as easily as possible? The ideal product would be a basket of offsets that's

  • easy to set up-- ideally a single monthly donation equivalent to the animal product consumption of the average American, which I can scale up a bit to make sure I'm net positive
  • based on well-founded impact estimates
  • affects a wide variety of animals reflecting my actual diet-- at a minimum my donation would be split among separate nonprofits improving the welfare of mammals, birds, fish, and invertebrates, and ideally it would closely track the suffering created by each animal product within that category
  • includes all animal products, not just meat.

I know I could potentially have higher impact just betting on saving 10 million shrimp or whatever, but I have enough moral uncertainty that I would highly value this kind of offset package. My guess is there are lots of people for whom going vegan is not possible or desirable, who would be in the same boat.

Suppose that the EA community were transported to the UK and US in 1776. How fast would slavery have been abolished? Recall that the slave trade ended in 1807 in the UK and 1808 in the US, and abolition happened between 1838-1843 in the British Empire and 1865 in the US.

Assumptions:

  • Not sure how to define "EA community", but some groups that should definitely be included are the entire staff of OpenPhil and CEA, anyone who dedicates their career choices or donates more than 10% along EA principles, and anyone with >5k EA forum karma.
  • EAs have the same proportion of the population as they do now, as well as the same relative levels of wealth, political power, intelligence, and drive.
  • EAs forget all our post-1776 historical knowledge, including the historical paths to abolition.
  • EA attention is split among other top causes of the day, like infectious disease and crop yields. I can't think of a reason why antislavery would be totally ignored by EAs though, as it seems huge in scope and highly morally salient to people like Bentham.
    • I'm also interested in speculating on other causes, I've just been thinking about abolition recently due to the 80k podcast with Prof. Christopher Brown.

Note that (according to ChatGPT) Quakers were more dedicated to abolition than EAs are to animal advocacy, have a much larger population, and deserve lots of moral credit for abolition in real life. But my guess would be that EAs could find some angles the Quakers wouldn't due to the consequentialist principles of EA. Maybe more evangelism and growth (Quaker population declined in the early 1800s), pragmatism about compensating slaveholders in the US as was done in the UK, or direct political action. Could EAs have gotten the Fugitive Slave Clause out of the Constitution?

It is not clear to me that EA branding is net positive for the movement overall or if it's been tarnished beyond repair by various scandals. Like, it might be that people should make a small personal sacrifice to be publicly EA, but it might also be that the pragmatic collective action is to completely rebrand and/or hope that EA provides a positive radical flank effect.

The reputation of EA at least in the news and on Twitter is pretty bad; something like 90% of the news articles mentioning EA are negative. I do not think it inherently compromises integrity to not publicly associate with EA even if you agree with most EA beliefs, because people who read opinion pieces will assume you agree with everything FTX did, or are a Luddite, or have some other strawman beliefs. I don't know whether EAF readers calling themselves EAs would make others' beliefs about their moral stances more or less accurate.

I don't think this is currently true, but if the rate of scandals continues, anyone holding on to the EA label would be suffering from the toxoplasma of rage, where the EA meme survives by sounding slightly good to the ingroup but extremely negative to anyone else. Therefore, as someone who is disillusioned with the EA community but not various principles, I need to see some data before owning any sort of EA affiliation, to know I'm not making some anti-useful sacrifice.

Given the Guardian piece, inviting Hannania to Manifest seems like an unforced error on the part of Manifold and possibly Lightcone. This does not change because the article was a hitpiece with many inaccuracies. I might have more to say later.

I want to slightly push back against this post in two ways:

  • I do not think longtermism is any sort of higher form of care or empathy. Many longtermist EAs are motivated by empathy, but they are also driven by a desire for philosophical consistency, beneficentrism and scope-sensitivity that is uncommon among the general public. Many are also not motivated by empathy-- I think empathy plays some role for me but is not the primary motivator? Cold utilitarianism is more important but not the primary motivator either [1]. I feel much more caring when I cook dinner for my friends than when I do CS research, and it is only because I internalize scope sensitivity more than >99% of people that I can turn empathy into any motivation whatsoever to work on longtermist projects. I think that for most longtermists, it is not more empathy, nor a better form of empathy, but the interaction of many normal (often non-empathy) altruistic motivators and other personality traits that makes them longtermists.
  • Longtermists make tradeoffs between other common values and helping vast future populations that most people disagree with, and without ideosyncratic EA values there is no reason that a caring person should make the same tradeoffs as longtermists. I think the EA value of "doing a lot more good matters a lot more" is really important, but it is still trading off against other values.
    • Helping people closer to you / in your community: many people think this has inherent value
    • Beneficentrism: most people think there is inherent value in being directly involved in helping people. Habitat for Humanity is extremely popular among caring and empathic people, and they would mostly not think it is better to make more of an overall difference by e.g. subsidizing eyeglasses in Bangladesh.
    • Justice: most people think it is more important to help one human trafficking victim than one tuberculosis victim or one victim of omnicidal AI if you create the same welfare, because they place inherent value on justice. Both longtermists and GiveWell think they're similarly good modulo secondary consequences and decision theory.
    • Discount rate, risk aversion, etc.: There is no reason that having a 10% chance of saving 100 lives in 6,000 years is better than a 40% chance of saving 5 lives tomorrow, if you don't already believe in zero-discount expected value as the metric to optimize. The reason to believe in zero-discount expected value is a thought experiment involving the veil of ignorance, or maybe the VNM theorem. It is not caring doing the work here because both can be very caring acts, it is your belief in the thought experiment connecting your caring to the expected value.

In conclusion, I think that while care and empathy can be an important motivator to longtermists, and it is valid for us to think of longtermist actions as the ultimate act of care, we are motivated by a conjunction of empathy/care and other attributes, and it is the other attributes that are by far more important. For someone who has empathy/care and values beneficentrism and scope-sensitivity, preventing an extinction-level pandemic is an important act of care; for someone like me or a utilitarian, pandemic prevention is also an important act. But for someone who values justice more, applying more care does not make them prioritize pandemic prevention over helping a sex trafficking victim, and in the larger altruistically-inclined population, I think a greater focus on care and empathy conflict with longtermist values more than they contribute.

[1] More important for me are: feeling moral obligation to make others' lives better rather than worse, wanting to do my best when it matters, wanting future glory and social status for producing so much utility.


Not sure how to post these two thoughts so I might as well combine them.

In an ideal world, SBF should have been sentenced to thousands of years in prison. This is partially due to the enormous harm done to both FTX depositors and EA, but mainly for basic deterrence reasons; a risk-neutral person will not mind 25 years in prison if the ex ante upside was becoming a trillionaire.

However, I also think many lessons from SBF's personal statements e.g. his interview on 80k are still as valid as ever. Just off the top of my head:

  • Startup-to-give as a high EV career path. Entrepreneurship is why we have OP and SFF! Perhaps also the importance of keeping as much equity as possible, although in the process one should not lie to investors or employees more than is standard.
  • Ambition and working really hard as success multipliers in entrepreneurship.
  • A career decision algorithm that includes doing a BOTEC and rejecting options that are 10x worse than others.
  • It is probably okay to work in an industry that is slightly bad for the world if you do lots of good by donating. [1] (But fraud is still bad, of course.)

Just because SBF stole billions of dollars does not mean he has fewer virtuous personality traits than the average person. He hits at least as many multipliers than the average reader of this forum. But importantly, maximization is perilous; some particular qualities like integrity and good decision-making are absolutely essential, and if you lack them your impact could be multiplied by minus 20.

 

 

[1] The unregulated nature of crypto may have allowed the FTX fraud, but things like the zero-sum zero-NPV nature of many cryptoassets, or its negative climate impacts, seem unrelated. Many industries are about this bad for the world, like HFT or some kinds of social media. I do not think people who criticized FTX on these grounds score many points. However, perhaps it was (weak) evidence towards FTX being willing to do harm in general for a perceived greater good, which is maybe plausible especially if Ben Delo also did market manipulation or otherwise acted immorally.

Also note that in the interview, SBF didn't claim his donations offset a negative direct impact; he said the impact was likely positive, which seems dubious.

Load more