# 681

We were shocked and immensely saddened to learn of the recent events at FTX. Our hearts go out to the thousands of FTX customers whose finances may have been jeopardized or destroyed.

We are now unable to perform our work or process grants, and we have fundamental questions about the legitimacy and integrity of the business operations that were funding the FTX Foundation and the Future Fund. As a result, we resigned earlier today.

We don’t yet have a full picture of what went wrong, and we are following the news online as it unfolds. But to the extent that the leadership of FTX may have engaged in deception or dishonesty, we condemn that behavior in the strongest possible terms. We believe that being a good actor in the world means striving to act with honesty and integrity.

We are devastated to say that it looks likely that there are many committed grants that the Future Fund will be unable to honor. We are so sorry that it has come to this. We are no longer employed by the Future Fund, but, in our personal capacities, we are exploring ways to help with this awful situation. We joined the Future Fund to support incredible people and projects, and this outcome is heartbreaking to us.

We appreciate the grantees' work to help build a better future, and we have been honored to support it. We're sorry that we won't be able to continue to do so going forward, and we deeply regret the difficult, painful, and stressful position that many of you are now in.

To reach us, grantees may email grantee-reachout@googlegroups.com. We know grantees must have many questions, and in our personal capacities we will try to answer them as best as we can given the circumstances.

Leopold Aschenbrenner

Avital Balwit

Ketan Ramakrishnan

# 681

New Comment
Some comments are truncated due to high volume. Change truncation settings

What do EA and the FTX Future Team think of a claim by Kerry Vaughan that Sam Bankman-Fried did severely unethical behavior before and EA and FTX covered it up and laundered his reputation, effectively getting away with it.

I'm posting because of true, this suggests big changes to EA norms are necessary to deal with bad actors like him, and that Sam Bankman-Fried should be outright banned from the forum and EA events.

I want to clarify the claims I'm making in the Twitter thread.

I am not claiming that EA leadership or members of the FTX Future fund knew Sam was engaging in fraudulent behavior while they were working at FTX Future Fund.

Instead, I am saying that friends of mine in the EA community worked at Alameda Research during the first 6 months of its existence. At the end of that period, many of them suddenly left all at once. In talking about this with people involved, my impression is:

1) The majority of staff at Alameda were unhappy with Sam's leadership of the company. Their concerns about Sam included concerns about him taking extreme and unnecessary risks and losing large amounts of money,  poor safeguards around moving money around, poor capital controls, including a lack of distinction between money owned by investors and money owned by Alameda itself, and Sam generally being extremely difficult to work with.

2) The legal ownership structure of Alameda did not reflect the ownership structure that had been agreed to by the parties involved.  In particular, Sam registered Alameda under his sole ownership and not as jointly owned by him and his cofounders. This was not thought t...

I was one of the people who left at the time described. I don't think this summary is accurate, particularly (3).

(1) seems the most true, but anyone who's heard Sam on a podcast could tell you he has an enormous appetite for risk. IIRC he's publicly stated they bet the entire company on FTX despite thinking it had a <20% chance of paying off. And yeah, when Sam plays league of legends while talking to famous investors he seems like a quirky billionaire; when he does it to you he seems like a dick. There are a lot of bad things I can say about Sam, but there's no elaborate conspiracy.

Lastly, my severance agreement didn't have a non-disparagement clause, and I'm pretty sure no one's did. I assume that you are not hearing from staff because they are worried about the looming shitstorm over FTX now, not some agreement from four years ago.

When said shitstorm dies down I might post more and under my real name, but for now the phrase "wireless mouse" should confirm me as someone who worked there at the time to anyone else who was also there.

I'm the person that Kerry was quoting here, and am at least one of the reasons he believed the others had signed agreements with non-disparagement clauses. I didn't sign a severance agreement for a few reasons: I wanted to retain the ability to sue, I believed there was a non-disparagement clause, and I didn't want to sign away rights to the ownership stake that I had been verbally told I would receive. Given that I didn't actually sign it, I could believe that the non-disparagement clauses were removed and I didn't know about it, and people have just been quiet for other reasons (of which there are certainly plenty).

I think point 3 is overstated but not fundamentally inaccurate. My understanding was that a group of senior leadership offered Sam to buy him out, he declined, and he bought them out instead. My further understanding is that his negotiating position was far stronger than it should have been due to him having sole legal ownership (which I was told he obtained in a way I think it is more than fair to describe as backstabbing). I wasn't personally involved in those negotiations, in part because I clashed with Sam probably worse than anyone else at the company, which likel...

I'm curious if you (or any other "SBF skeptic") has any opinion regarding whether his character flaws should've been apparent to more people outside the organizations he worked at, e.g. on the basis of his public interviews. Or alternatively, were there any red flags in retrospect when you first met him?

I'm asking because so far this thread has discussed the problem in terms of private info not propagating. But I want to understand if the problem could've been stopped at the level of public info. If so that suggests that a solution of just getting better at propagating private info may be unsatisfactory -- lots of EAs had public info about SBF, but few made a stink.

I'm also interested to hear "SBF skeptic" takes on the extent his character flaws were a result of his involvement in EA. Or maybe something about being raised consequentialist as a kid? Like, if we believe that SBF would've been a good person if it weren't for exposure to consequentialist ideas, that suggests we should do major introspection.

One of the biggest lessons I learned from all of this is that while humans are quite good judges of character in general, we do a lot worse in the presence of sufficient charisma, and in those cases we can't trust our guts, even when they're usually right. When I first met SBF, I liked him quite a bit, and I didn't notice any red flags. Even during the first month or two of working with him, I kind of had blinders on and made excuses for things that in retrospect I shouldn't have.

It's hard for me to say about what people should have been able to detect from his public presence, because I haven't watched any of his public interviews. I put a fair amount of effort into making sure that news about him (or FTX) didn't show up in any of my feeds, because when it did I found it pretty triggering.

Personally, I don't think his character flaws are at all a function of EA. To me, his character seems a lot more like what I hear from friends who work in politics about what some people are like in that domain. Given his family is very involved in politics, that connection seems plausible to me. This is very uncharitable, but: from my discussions with him he always seemed a lot more interested in power than in doing good, and I always worried that he just saw doing good as an opportunity to gain power. There's obviously no way for me to have any kind of confidence in that assessment, though, and I don't think people should put hardly any weight on it.

In terms of public interviews, I think the most interesting/relevant parts are him expressing willingness to bite consequentialist/utilitarian bullets in a way that's a bit on the edge of the mainstream Overton window, but I believe would've been within the EA Overton window prior to recent events (unsure about now). BTW I got these examples from Marginal Revolution comments/Twitter.

• This one seems most relevant -- the first question Patrick asks Sam is whether the ends justify the means.

• In this interview, search for "So why then should we ever spend a whole lot of money on life extension since we can just replace people pretty cheaply?" and "Should a Benthamite be risk-neutral with regard to social welfare?"

In any case, given that you think people should put hardly any weight on your assessment, it seems to me that as a community we should be doing a fair amount of introspection. Here are some things I've been thinking about:

• We should update away from "EA exceptionalism" and towards self-doubt. (EDIT: I like this thread about "EA exceptionalism", though I don't agree with all the claims.) It sounds like you think more self-doubt would've been re

...

This one is tricky, because it seems bad to tell people who already experience Chidi Anagonye-style crippling self-doubt that they should self-doubt even more.

EA self-doubt has always seemed weirdly compartmentalized to me. Even the humblest of people in the movement is often happy to dismiss considered viewpoints by highly intelligent people on the grounds that it doesn't satisfy EA principles. This includes me - I think we are sometimes right to do so, but probably do so far too much nonetheless.

1John_Maxwell21d
Seems plausible, I think it would be good to have a dedicated "translator" who tries to understand & steelman views that are less mainstream in EA. Wasn't sure about the relevance of that link?

(from phone) That was an example of an ea being highly upvoted for dismissing multiple extremely smart and well meaning people's life's work as 'really flimsy and incredibly speculative' because he wasn't satisfied that they could justify their work within a framework that the ea movement had decided is one of the only ones worth contemplating. As if that framework itself isn't incredibly speculative (and therefore if you reject any of its many suppositions, really flimsy)

7John_Maxwell20d
Thanks! I'm not sure I share your view of that post. Some quotes from it: ... ... ... I don't think any of these observations hinge on the EA framework strongly? Like, do we have reason to believe Andrew Carnegie spent a significant amount trying to figure out if libraries were a great donation target by his own lights , as opposed to according to the EA framework? The thing that annoyed me about that post was that at the time it was written, it seemed to me that the EA movement was also fairly guilty of this! (It was written before the criticism/red teaming contest.)
7Arepo20d
I'm not familiar enough with the case of Andrew Carnegie to comment and I agree on the point of political tribalism. The other two are what bother me. On the professor, the problem is there explicitly: you omitted a key line 'I tried asking for his opinion on existential threats', which is a strongly EA-identifying approach, and one which many people feel is too simplistic. Eg see Gideon Futurman's EAGx Rotterdam talk when it's up - he argues the way EAs think about x-risk is far too simplified, focusing on single-event narratives, ignoring countless possible trajectories that could end in extinction or similar any one of which is vanishingly unlikely, but which collectively we should take much more seriously. Whether or not one agrees with this view, it seems to me to be one a smart person could reasonably hold, and shows that by asking someone 'his opinion on existential threats, and which specific scenarios these space settlements would help with', you're pigeonholing them into EA-aligned specific-single-event way of thinking. As for Elon Musk, I think the same problem is there implicitly: he's written a paper called 'Making Humans a Multiplanetary Species', spoken extensively on the subject and spent his life thinking that it's important, and while you could reasonably disagree with his arguments, I don't see any grounds for dismissing them as 'really flimsy and incredibly speculative' without engagement, unless your reason for doing so is 'there exists a pool of important research which contradicts them and which I think is correct'. There are certainly plenty of other smart people who think as he does, some of them EAs [https://forum.effectivealtruism.org/topics/space-governance] (though maybe that doesn't contribute to my original complaint). Since there's a very clear mathematical argument that it's harder to kill all of a more widespread and numerous civilisation, to say that the case is 'really flimsy', you basically need to assume the EA-aligned narrati
4John_Maxwell19d
Thanks!
3Simon Bazelon21d
What's interesting about this interview clip though is that he seems to explicitly endorse a set of principles that directly contradict the actions he took!
4John_Maxwell21d
Well that's the thing -- it seems likely he didn't see his actions as contradicting those principles. Suggesting that they're actually a dangerous set of principles to endorse, even if they sound reasonable. That's what's really got me thinking. I wonder if part of the problem is a consistent failure of imagination on the part of humans to see how our designs might fail. Kind of like how an amateur chess player devotes a lot more thought to how they could win than how their opponent could win. So if the principles Sam endorsed are at all recoverable, maybe they could be recovered via a process like "before violating common-sense ethics for the sake of utility, go down a massive checklist searching for reasons why this could be a mistake, including external observers in the decision if possible".
3Sharmake21d
My guess is standard motivated reasoning explains why he thought he wasn't in violation of his stated principles. Question, but why do you think the principles were dangerous, exactly? I am confused about the danger you state.
4John_Maxwell21d
I think your first paragraph provides a potential answer to your second :-) There's an implicit "Sam fell prey to motivated reasoning, but I wouldn't do that" in your comment, which itself seems like motivated reasoning :-) (At least, it seems like motivated reasoning in the absence of a strong story for Sam being different from the rest of us. That's why I'm so interested in what people like nbouscal have to say.)
4Sharmake21d
So you think there's too much danger of cutting yourself and everyone else via motivated reasoning, ala Dan Luu's "Normalization of Deviance" and the principles have little room for errors in implementing them, is that right? Here's a link to it: https://danluu.com/wat/ [https://danluu.com/wat/] And a quote:
3John_Maxwell21d
I'm not sure what you mean by "the principles have little room for errors in implementing them". That quote seems scarily plausible. EDIT: Relevant Twitter thread [https://twitter.com/hamandcheese/status/1590841627979153409]
3Sharmake21d
Specifically, I was saying that wrong results would come up if you failed in one of the steps of reasoning, and there's no self-correction mechanism for bad reasoning like Sam Bankman-Fried was doing.

I do feel that my perspective has been clearly vindicated by current events.

Can I ask the obvious question of whether you made money by shorting ftt? You were both one of the most anti-FTX and most still involved in crypto trading, so I suspect if you didn't then no one did.

Ps: apologies for burning the "wireless mouse" Commons. If others want to make throwaways, feel free to dm me what that is referring to and I will publicly comment my verification.

Also no non-disparagement clause in my agreement. FWIW I was one of the people who negotiated the severance stuff after the 2018 blowup, and I feel fairly confident that that holds for everyone. (But my memory is crappy, so that's mostly because I trust the FB post about what was negotiated more than you do.)

DM'd you.

Confirming this account made an Alameda research reference in my DMs.

5nbouscal20d
… I assume you realise that that narrows you down to one of two people (given it's safe to assume Nishad is not currently spending his time on the EA Forum) I do think I was probably just remembering incorrectly about this to be honest, I looked back through things from then and it looks like there was a lot of back-and-forth about the inclusion of an NDA (among other clauses), so it seems very plausible that it was just removed entirely during that negotiation (aside from the one in the IP agreement).
4arthrowaway20d
0Anonymous (for unimpressive reasons =[ )20d
Here is some questions/content [https://forum.effectivealtruism.org/posts/xafpj3on76uRDoBja/the-ftx-future-fund-team-has-resigned-1?commentId=fo2jDE4uFHhxZ6rHB] that might be interesting to discuss if you're interested?

Can I ask the obvious question of whether you made money by shorting ftt? You were both one of the most anti-FTX and most still involved in crypto trading, so I suspect if you didn't then no one did.

I've been on leave from work due to severe burnout for the last couple months (and still am), and was intentionally avoiding seeing anything about SBF/FTX outside of work until recent events made that basically impossible. So no, I didn't personally trade on any of this at all.

Fair. Sorry to hear that, I hope you can go back to ignoring the situation soon!

9Anonymous (for unimpressive reasons =[ )20d
Can you answer two questions related to the source of SBF's early business wealth? Were the Kimchi arb returns real? As you know, the "Kimchi premium" was this difference in BTC price between Korea (Japan?) and the rest of the world. The narrative is that SBF arbed this price difference to make many millions and create his early wealth. The Sequoia puff piece [https://web.archive.org/web/20221027181005/https://www.sequoiacap.com/article/sam-bankman-fried-spotlight/] makes this cute story: After SBF's fall, Twitter speculation says this is dubious. This is because the cause of the Kimchi premium was strict legal capital controls, and the liquidity was orders of magnitude too small to produce the wealth in SBF later used. At best, SBF was actively breaking laws by this trade. The amount of money he could make may have been too small to justify the narratives around his early success. Do you have any comments on the above? Jaan Tallinn investment Tallinn later ended up funding SBF with 50M. What would you say to the speculation that it was this funding, and not the Kimchi arb , that really launched SBF's career? If this is mostly true, the takeaway is that there's little cleverness or competency being expressed here here? It seems like power, money and access led to SBF's success. This theme would fit with SBF's later behavior, with bluffing and overaweing spend. That tradition seems hollow and bad, maybe contagious to the things that SBF created or touched. This could be useful in some way? It seems like the vector EA or EA PR could take, could counter this. I don't mind sharing a bit about this. SBF desperately wanted to do the Korea arb, and we spent quite a bit of time coming up with any number of outlandish tactics that might enable us to do so, but we were never able to actually figure it out. The capital controls worked. The best we could do was predict which direction the premium would go and trade into KRW and then back out of it accordingly. Japan was different. We were able to get a Japanese entity set up, and we did successfully trade on the Japan arb. As far as I know we didn't break any laws in doing so, but I wasn't directly involved in the operational side of it. My recollection is that we made something like 10-30 million dollars (~90%CI) off of that arb in total, but I'm not at all confident on the exact amount. Is that what created his early wealth, though? Not really. Before we all left, pretty much all of that profit had been lost to a series of bad trades and mismanagement of assets. Examples included some number of millions lost to a large directional bet on ETH (that Sam made directly counter to the predictions of our best event trader), a few million more on a large OTC trade in some illiquid shitcoin that crashed... 4Sabs20d What about the GBTC arb trade? Did Alameda get into that during your time there? 4nbouscal20d Good question, but tbh I just don’t remember the answer. 4Yitz20d Thank you for sharing, I can understand why you might be feeling burnt out!! I've been in a workplace environment that reminds me of this, and especially if you care about the people and projects there...it's painful. -1Anonymous (for unimpressive reasons =[ )20d Here is some questions/content [https://forum.effectivealtruism.org/posts/xafpj3on76uRDoBja/the-ftx-future-fund-team-has-resigned-1?commentId=fo2jDE4uFHhxZ6rHB] that might be interesting to discuss? (You might not want to given if your fatigue though.) Thanks for sharing this nbouscal. How many people did you tell about this at the time? Personally, I remember telling at least a handful of people at the time that Sam belonged in a jail cell, but I expect that people thought I was being hyperbolic (which was entirely fair, I was traumatised and was probably communicating in a way that signalled unreliability). I was told that conversations were had with people in leadership roles in EA. I wasn’t part of those conversations and don’t know the full details of what was discussed or with whom. It would be awesome for the names of senior people who knew to be made public, plus the exact nature of what they were told and their response or lack thereof. 5Ozzie Gooen21d I think this could be a nice-to-have, but really, I think it's too much to ask, "For every senior EA, we want a long list of exactly each thing they knew about SBF" This would probably be a massive pain, and much of the key information will be confidential (for example, informants who want to remain anonymous). My guess is that there were a bunch of flags that were more apparent than nbouscal's stories. I do think we should have really useful summaries of the key results. If there were a few people who were complicit or highly negligent, then that should be reported, and appropriate actions taken. 9Devon Fritz21d I strongly believe it is hyperrelevant to know who knew what, when so that these people are held to account. I don't think this is too much to ask, nor does it have to be arduous in the way you described of getting every name with max fidelity. I see so many claims that "key EA members knew what was going on" and never any sort of name associate with it. 8Ozzie Gooen21d I agree this is really important and would really, really want it to be figured out, and key actions taken. I think I'm less focused on all of the information of such a discovery being public, as opposed to much of it being summarized a bit. 2Isaac King19d A summary of sorts is being compiled here: 2pseudonym21d What would you suggest might be appropriate actions for complicity or negligence? 2Ozzie Gooen21d I don't feel like I'm in a good place to give a good answer. First, I haven't really thought about it nor am I an expert in these sorts of matters. Second, I'm like several layers deep in funding structures that start with these people. It's sort of like asking me to publicly write what I love/hate, objectively, about my boss. I think I could say that I'd expect appropriate actions to look a lot like they do with top companies (mainly ones without lots of known management integrity problems). At these companies, I believe that when some officials are investigated for potential issues, often they're given no punishment, and sometimes they're fired. It really depends on the details of the findings. I think it is very important to understand what was known about SBF's behaviour during the initial Alameda breakup, and for this to be publicly discussed and to understand if any of this disaster was predictable beforehand. I have recently spoken to someone involved who told me that SBF was not just cavalier, but unethical and violated commonsense ethical norms. We really need to understand whether this was known beforehand, and if so learn some very hard lessons. It is important to distinguish different types of risk-taking here. (1) There is the kind of risk taking that promises high payoffs but with a high chance of the bet falling to zero, without violating commonsense ethical norms, (2) Risk taking in the sense of being willing to risk it all secretly violating ethical norms to get more money. One flaw in SBF's thinking seemed to be that risk-neutral altruists should take big risks because the returns can only fall to zero. In fact, the returns can go negative - eg all the people he has stiffed, and all of the damage he has done to EA. I have recently spoken to someone involved who told me that SBF was not just cavalier, but unethical and violated commonsense ethical norms. Are you in a position to be more specific about what SBF did that this is referring to? 8John G. Halstead22d no In 2021 I tried asking about SBF among what I suppose you could call "EA leadership", trying to distinguish whether to put SBF into the column of "keeps compacts but compact very carefully" versus "un-Lawful oathbreaker", based on having heard that early Alameda was a hard breakup. I did not get a neatly itemized list resembling this one on either points 1 or 2, just heard back basically "yeah early Alameda was a hard breakup and the ones who left think they got screwed" (but not that there'd been a compact that got broken) (and definitely not that they'd had poor capital controls), and I tentatively put SBF into column 1. If "EA leadership" had common knowledge of what you list under items 1 or 2, they didn't tell me about it when I asked. I suppose in principle that I could've expended some of my limited time and stamina to go and inquire directly among the breakup victims looking for one who hadn't signed an NDA, but that's just a folly of perfect hindsight. My own guess is that you are mischaracterizing what EA leadership knew. Huh, I am surprised that no one responded to you on this. I wonder whether I was part of that conversation, and if so, I would be interested in digging into what went wrong. I definitely would have put Sam into the "un-lawful oathbreaker" category and have warned many people I have been working with that Sam has a reputation for dishonesty and that we should limit our engagement with him (and more broadly I have been complaining about an erosion of honesty norms among EA leadership to many of the current leadership, in which I often brought up Sam as one of the sources of my concern directly). I definitely had many conversations with people in "EA leadership" (which is not an amazingly well-defined category) where people told me that I should not trust him. To be clear, nobody I talked to expected wide-scale fraud, and I don't think this included literally everyone, but almost everyone I talked to told me that I should assume that Sam lies substantially more than population-level baseline (while also being substantially more strategic about his lying than almost everyone else). I do want to add to this that in addition to Sam having a reputation for dishonesty, he also had a reputation for being vindictive, and almost everyone who told me about their concerns about Sam did so while seeming quite visibly afraid of retribution from Sam if they were to be identified as the source of the reputation, and I was never given details without also being asked for confidentiality. Can you give some context on why Lightcone accepted a FTX Future Fund grant (a) given your view of his trustworthiness? So far I have been running on the policy that I will accept money from people who seem immoral to me, and indeed I preferred getting money from Sam instead of Open Philanthropy or other EA funders because I thought this would leave the other funders with more marginal resources that could be used to better ends (Edit: I also separately thought that FTX Foundation money would come with more freedom for Lightcone to pursue its aims independently, which I do think was a major consideration I don't want to elide). To be clear, I think there is a reasonable case to be made for the other end of this tradeoff, but I currently still believe that it's OK for EAs to take money from people whose values or virtues they think are bad (and that indeed this is often better than taking money from the people who share your values and virtues, as long as its openly and willingly given). I think the actual tradeoffs are messy, and indeed I ended up encouraging us to go with a different funder for a loan arrangement for a property purchase we ended up making, since that kind of long-term relationship seemed much worse to me, and I was more worried about that entangling us more with FTX. To b... In situations like this, it might be a good habit to state reservations publicly at the same time you receive the grant? Then your accepting the grant isn't a signal that you endorse the grantmaker, and you can be less worried about your relationship with the grantmaker damaging your future ability to be candid. Either they stop giving you money, or they continue giving you money even though you badmouthed them (which makes it more clear that you have impunity to do so again in the future). 3Geoffrey Miller19d Interesting idea. But it seems unrealistic to expect a recipient of a grant, upon receiving it, to publicly announce ethical and legal reservations about the grant-giver... and then for the grant-giver to be OK with that, and to follow through on providing the grant funding. 'Biting the hand that feeds you' doesn't typically result in good outcomes. Sure, though I think altruistic grantmakers should want their grantees to criticize them (because an altruistic grantmaker should care more about getting useful and actionable criticism than about looking good in the moment), and I think a lot of EA grantmakers walk the walk in that respect. E.g., MIRI has written tons of stuff publicly criticizing Open Phil, even though Open Phil is by far our largest all-time funder; and I don't think this has reduced our probability of getting future Open Phil funding. One advantage of the norm I proposed is that it can help make this a more normal and expected practice, and (for that reason) less risky than it currently is. And since everything's happening in public, grantmakers can accumulate track records. If you keep defunding people when they criticize you (even when the criticisms seem good and the grant recipients seem worthy, as far as others can tell), others can notice this fact and dock the grantmaker reputational points. (Which should matter to grantmakers who are optimizing this hard for their reputation in the first place.) Fair points. I guess if any community can create a norm where it's OK for grant receivers to criticize grantmakers, it's the EA community. I was really just pointing out that creating and maintaining such an open, radically honest, self-reflective, criticism-welcoming culture is very much an uphill struggle, given human nature. That's very surprising!! Do you know if anybody attempted to propagate this information to any of the EAs who were promoting SBF publicly? (If so, do you know if they succeeded in conveying that information to them?) And just to check, did any of the people who warn you privately promote SBF/FTX publicly? I ask because it seems weird for a lot of EAs to be passing around warnings about SBF being untrustworthy while a lot of (other?) EAs are promoting him publicly; I very much hope these sets were disjoint, but also it's weird for them to be so disjoint, I would have expected better information flow. Yep, I was and continue to be confused about this. I did tell a bunch of people that I think promoting SBF publicly was bad, and e.g. sent a number of messages when some news article that people were promoting (or maybe 80k interview?) was saying that "Sam sleeps on a bean bag" and "Sam drives a Corolla" when I was quite confident that they knew that Sam was living in one of the most expensive and lavish properties in the Bahamas and was definitely not living a very frugal livestyle. This was just at the same time as the Carrick stuff was happening, and I would have likely reached out to more people if I hadn't spent a lot of my social energy on pushing back on Carrick stuff at the time (e.g. ASB's piece on Carrick's character). Overall, I did not message many people, and I personally did not speak out very publicly about my broader concerns. I also think a lot of that promotion was happening in a part of the EA ecosystem I interface much less with (80k, UK EAs, Will, etc.), and I've had historically somewhat tense relationships to that part of the ecosystem, so I did not have many opportunities to express my concerns. It would be useful to say whether any of the people you told would be considered 'EA leadership'; and if so, who. How can both of these be true: 1. You (and others, if all of the accounts I've been reading about are true) told EA leadership about a deep mistrust of SBF. 2. EA decided to hold up and promote SBF as a paragon of EA values and on of the few prominent faces in the EA community. If both of those are true, how many logical possibilities are there? 1. The accounts that people told EA leadership are false. 2. The accounts are true and EA leadership didn't take these accounts seriously. 3. EA leadership took the accounts seriously, but still proceeded to market SBF. I find them all super implausible so I don't know what to think! My understanding is that the answer is basically 2. I'd love to share more details but I haven't gotten consent from the person who told me about those conversations yet, and even if I were willing to share without consent I'm not confident enough of my recollection of the details I was told about those conversations when they happened to pass that recollection along. I hope to be able to say more soon. EDIT: I've gotten a response and that person would prefer me not to go into more specifics currently, so I'm going to respect that. I do understand the frustration with all of the vagueness. I'm very hopeful that the EA leaders who were told about all of this will voluntarily come forward about that fact in the coming days. If they don't, I can promise that they will be publicly named eventually. My guess is different parts of leadership. I don't think many of the people I talked to promoted SBF a lot. E.g. see my earlier paragraph on a lot of this promotion being done by the more UK focused branches that I talk much less to. 4Devon Fritz20d That could very well be and there are a lot of moving parts. That is why I think it is important for people who supposedly warned leadership to say who was told and what they were told. If we are going to unravel this this all feels like necessary information. The people who are staying quiet about who they told have carefully considered reasons for doing so, and I'd encourage people to try to respect that, even if it's hard to understand from outside. My hope is that the information will be made public from the other side. EA leaders who were told details about the events at early Alameda know exactly who they are, and they can volunteer that information at any time. It will be made public eventually one way or another. 5Devon Fritz20d I respect that people who aren't saying what they know have carefully considered reasons for doing so. I am not confident it will come from the other side as it hasn't to date and there is no incentive to do so. May I ask why you believe it will be made public eventually? I truly hope that is the case. The incentives for them to do so include 1) modelling healthy transparency norms, 2) avoiding looking bad when it comes out anyway, 3) just generally doing the right thing. I personally commit to making my knowledge about it public within a year. (I could probably commit to a shorter time frame than that, that's just what I'm sure I'm happy to commit to having given it only a moment's thought.) What do you find super implausible about 2? 0Devon Fritz21d If insiders were making serious accusations about his character to EA leadership and they went on to promote him that would be weird to me. Especially if many people did it which is what has been claimed. Of course I have no idea who “leadership” is because no one is being specific. To be fair sometimes people make accusations that are incorrect? Your decision procedure does need to allow for the possibility of not taking a given accusation seriously. I don't know who knew what and how reasonable a conclusion this was for any given person given their state of knowledge, in this case, but also people do get this wrong sometimes, this doesn't seem implausible to me. 2Devon Fritz20d My decision procedure does allow for that and I have lots of uncertainties, but it feels that given many insiders claim to have warned people in positions of power about this and Sam got actively promoted anyway. If multiple people with intimate knowledge of someone came to you and told you that they thought person X was of bad character you wouldn't have to believe them hook line and sinker to be judicious about promoting that person. Maybe this is the most plausible of the 3 and I shouldn't have called it super implausible, but it doesn't seem very plausible for me, especially from people in a movement that takes risks more seriously than any other that I know. I found this comment annoying enough to read that I felt compelled to give a simplified version: In 2021, I asked about SBF among some "senior EA people". I had heard that Alameda had had a hard breakup, and I didn't know whether SBF cheated his partners or whether he was merely a punctilious negotiator who nonetheless keeps his word. Based on what I heard, I classified SBF more as a punctilious negotiator than as a cheater, and I definitely didn't hear about them having poor capital controls at the beginning. My own guess is that you are mischaracterizing what EA leadership knew. This removes some nuance, but maybe adds some clarity. Edit: Reworded, see original here. I did not say that it'd be good if somebody was a ruthless negotiator. If you're going to paraphrase somebody, please be more careful to paraphrase things they actually said, by dereferencing, and not add implications you thought they meant. 4NunoSempere21d I didn't say I was paraphrasing you, I said I was giving a simplified version. I also pointed out the sentence was not in the original. 7interstice21d Adding in an unflattering sentiment that was not said or clearly implied in the original is not "simplifying". 3NunoSempere20d Ok, fine, reworded. You can still find the original here [https://gist.github.com/NunoSempere/713e83f1dd08f9c17a03e87b84b6ebe1]. I consider this credible. It suggests that my categorization of "EA leadership" was probably too broad and that fewer people knew the details of the situation than I believed. That means there is a question of how many people knew. I am confident that Nick Beckstead and Will MacAskill knew about the broken agreement and other problems at Alameda. I am confident they are not the only ones that knew. Why are you confident of that? In general, I think there's just less time and competence and careful checking to go around, in this world, than people would want to believe. This isn't Hieronym's To The Stars or the partially Hieronym-inspired world of dath ilan. Huge thanks for spelling out the specific allegations about SBF's behavior in early Alameda; for the past couple days I'd been seeing a lot of "there was known sketchy stuff at Alameda in 2017-18" and it was kind of frustrating how hard it was to get any information about what is actually alleged to have happened, so I really appreciate this clear point-by-point summary. 2Yitz20d Same here, this is really helping me understand the (at least perceived) narrative flow of events 7Sharmake22d I decided to speak about it because if true, it would imply bad things about how EA hasn't remembered the last time things went wrong. In many senses, this is EA's first adversarial interaction, where we can't rely on internal norms of always cooperating anymore. 4Jason22d After the involved EAs consult with their lawyers, they may find a receptive audience to tell their stories at the Department of Justice or another federal agency. I would be shocked if the NDAs were effective as against cooperating with a federal investigation. If the quoted description is true, it seems relevant to the defense SBF seems to be trying to set up. I knew about Sam's bad character early on, and honestly I'm confused about what people would have expected me to do. I should have told people that Sam has a bad character and can't be trusted, that FTX is risky? Well, I did those things, and as far as I can tell, that has made the current situation less bad than it would have been otherwise (yes, it could have been worse!). In hindsight I should have done more of this though. Should I have told the authorities that Sam might be committing fraud? All I had were vague suspicions about his character and hints that he might be dishonest, but no convincing evidence or specific worries about fraud. (Add jurisdictional problems, concerns about the competence of regulators, etc) Should I not have "covered up" the early scandal? Well, EAs didn't, and I think Kerry's claim is wrong. Should I have publicly spread concerns about SBF's character? That borders on slander. Also, I was concerned that SBF would permanently hate me after that (you might say I'm a coward, but hey, try it yourself). Should I have had SBF banned from EA? Personally, I'm all for a tough stance, but the community is usually against complete bans of bad actors, so it just wasn't feasible. (EG, if I were in charge, Jacy and Kerry would be banned, but many wouldn't like that.) SBF was powerful and influential. EA didn't really have power over him. What could have been done better? I am sincerely curious to get suggestions. My current, extremely tentative, sense of the situation is not that individuals who were aware of some level of dishonesty and shadiness were not open enough about it. I think individuals acted in pretty reasonable ways, and I heard a good amount of rumors. I think the error likely happened at two other junctions: 1. Some part of EA leadership ended up endorsing SBF very publicly and very strongly despite having very likely heard about the concerns, and without following up on them (In my model of the world Will fucked up really hard here) 2. We didn't have any good system for aggregating rumors and related information, and we didn't have anyone who was willing to just make a public post about the rumors (I think this would have been a scary and heroic thing to do, I am personally ashamed that I didn't do it, but I don't think it's something that we should expect the average person to do) I think if we had some kind of e.g. EA newspaper where people try to actively investigate various things that seem concerning, then I think this would have helped a bunch. This kind of thing could even be circulated privately, though a public version seems also good. I separately also think... Agree with much of what you say here. (Though I don't think we currently have strong enough evidence to single out specific EA leaders as being especially responsible for the recent tragic events; at least I don't think I personally have that kind of information.) As a substitute, or complement, to an investigative EA newspaper, what do you think about an "EA rumours" prediction market?[1] Some attractive features of such a market: • It turns private information held by individual people with privileged access to sources into public information available to the entire EA community, increasing the likelihood that the information will reach those for whom it is most valuable and actionable. • It potentially reduces community drama by turning "hot" debates influenced by tribal allegiances and virtue signaling into "cold" assignments of probability and assessments of evidence. • It makes rumours more accurate, by incentivizing users to estimate their probability correctly. • It makes false rumours less damaging to their targets, by explicitly associating them with a low probability. I think this market would need judicious moderation to function well and avoid being abused. But overall it seems... Yeah, I feel excited about something in this space. Generally I feel like prediction markets have a lot of good things going for them in situations like this, though I do worry that they will somehow just end up gamed when the stakes are high. Like, my guess is Sam could have likely moved the probability of a market here a lot, either with money, or by encouraging other people to move it. Confirming that this account DM'd me with information indicating that they worked at Alameda. 9Ben Snodin19d Should EA people just be way more aggressive about spreading the word (within the community, either publicly or privately) about suspicions that particular people in the community have bad character? (not saying that this is an original suggestion, you basically mention this in your thoughts on what you could have done differently) 4ZekeFaux19d I met Sam in February and wrote a profile of him for Bloomberg [https://www.bloomberg.com/news/features/2022-04-03/sam-bankman-fried-ftx-s-crypto-billionaire-who-wants-to-give-his-fortune-away?sref=sGNHChN7] . In hindsight, there are a lot of red flags that everyone missed, myself included. Of course, it all looked different when he was on top. At the time, I tried to research Alameda's early years and the dispute that led to the big breakup, but didn't get anywhere. I'm now working on a book - on the off chance that any insiders from Alameda or FTX read this, please DM me here or on Twitter. I'm unclear how to update on this, but note that Kerry Vaughan was at CEA for 4 years, as a managing director for one before, as I understand it, being let go under mysterious circumstances. He's now the program manager at a known cult that the EA movement has actively distanced itself from. So while his comments are interesting, I wouldn't treat him as a particularly credible source, and he may have his own axe to grind. All this conversation about Leverage and Kerry's motives and character misses the point that he's talking about events that have little to nothing to do with him. He's saying that there was a blowup at Alameda early on reflecting badly on SBF that lots of EA leaders knew about and turned a blind eye to. This can be investigated and confirmed or denied without delving into conversations about Leverage or Kerry that are besides the point at hand. To the extent that Kerry's allegation involves his own judgment of Sam's actions as bad or shady, I think it matters that there's reason not to trust Kerry's judgment or possibly motives in sharing the information. However we should definitely try to find out what actually happened and determine whether it was truly predictive of worse behavior down the line. Agreed! IMO it's good for people to be aware that Kerry has an axe to grind; but the thing to do with that information is to look into the matter further. I commented above that I think Kerry's comment is incorrect, so I feel obligated to state that I have no reason to think this is the result of bias. I am inclined to think he's doing the best he can in an information-scarce environment. I retract this comment. Kerry has continued repeating the same claim on Twitter without noting that there's disagreement about its truth. This does not seem like unbiased behavior. 5Kerry_Vaughan20d The claim on Twitter is different. Can you clarify what you think is unfair? Happy to issue a correction. https://twitter.com/KerryLVaughan/status/1591508739236188160?t=qL-dGKXar3b7EQ4EHs597Q&s=19 [https://twitter.com/KerryLVaughan/status/1591508739236188160?t=qL-dGKXar3b7EQ4EHs597Q&s=19] Edit: if anyone else wants to take a stab at explaining why the Twitter thread is unfair given this thread feel free. Would want to issue a correction sooner rather than later. -1 on this comment. In particular, being at CEA for 4 years seems like something which makes criticism more plausible. And it's not surprising that EA has distanced itself from groups critical of us (while I have some concerns about Leverage, I think there are a bunch of ways that they've been treated unfairly). Hard disagree on Leverage. They've absorbed a tonne of philanthropic funding over the years to produce nothing but pseudoscience and multiple allegations of emotional abuse. I'm not saying Kerry wouldn't know about this stuff - I think he likely does. I'm saying a) that he was one of the 'top leaders' he refers to, so had ample chance to do something about this himself, b) he has a track record of questionable integrity, and c) he has potential motive to undermine the people he's criticising. I think this comment is a pretty clear example of one way in which Leverage has been treated unfairly, which is that people lump "not very productive" and "abusive" into a single criticism. The latter is much more serious, but the former is much easier to quickly verify, and so the former ends up lending credibility to the latter even though I personally think we probably have too few groups taking philanthropic funding to do crazy research that may end up looking like pseudoscience. To be very clear, I'm not claiming that Leverage was not an abusive environment, and I take the allegations you mention very seriously. I've just also seen people piling onto Leverage in not-very-careful ways that I'm not very happy about. I'm not a fan of Leverage, but I agree with Richard here. I think Kerry is better modeled as "normal philosophy-friendly EA" with the modifications "less conflict-averse than the average EA" and "mad at EA (for plenty of good reasons and also plenty of bad reasons, IMO) and therefore pretty axe-grindy". If you model him with a schema closer to "crazy cultist" than to "bitter ex-EA", I expect you to make worse predictions. I’m guessing I have a lower opinion of Leverage than you based on your tone, but +1 on Kerry being at CEA for 4 years making it more important to pay serious attention to what he has to say even if it ultimately doesn’t check out. We need to be very careful to minimize tribalism hurting our epistemics. For what it's worth, these different considerations can be true at the same time: 1. "He may have his own axe to grind.": that's probably true, given that he's been fired by CEA. 2. "Kerry being at CEA for four years makes it more important to pay serious attention to what he has to say even if it ultimately doesn’t check out.": it also seems like he may have particularly useful information and contexts. 3. "He's now the program manager at a known cult that the EA movement has actively distanced itself from": it does seem like Leverage is shady and doesn't have a very good culture and epistemic, which doesn't reflect greatly on Kerry. So I would personally be inclined to pay close attention to his criticisms of CEA. At the same time, I would need more "positive" contexts from others to be able to trust what he says. I agree that these can technically all be true at the same time, but I think the tone/vibe of comments is very important in addition to what they literally say, and the vibe of Arepo's comment was too tribalistic. I'd also guess re: (3) that I have less trust in CEA's epistemics to necessarily be that much better than Leverage's , though I'm uncertain here (edited to add: tbc my best guess is it's better, but I'm not sure what my prior should be if there's a "he said / she said" situation, on who's telling the truth. My guess is closer to 50/50 than 95/5 in log odds at least). I agree that the tone was too tribalistic, but the content is correct. (Seems a bit like a side-topic, but you can read more about Leverage on this EA Forum post and, even more importantly, in the comments. I hope that's useful for you! The comments definitely changed my views - negatively - about the utility of Leverage's outputs and some cultural issues.) I've read it. I'd guess we have similar views on Leverage, but different views on CEA. I think it's very easy for well-intentioned, generally reasonable people's epistemics to be corrupted via tribalism, motivated reasoning, etc. But as I said above I'm unsure. Edited to add: Either way, might be a distraction to debate this sort of thing further. I'd guess that we both agree in practice that the allegations should be taken seriously and investigated carefully, ideally by independent parties. Mea culpa for not being clear enough. I don't think handwavey statements from someone whose credibility I doubt have much evidential value, but I strongly think CEA's epistemics and involvement should be investigated - possibly including Vaughan's. I find it bleakly humourous to be interpreted as tribalistically defending CEA when I've written gradually more public criticisms of them and their lack of focus -and honestly, while I don't understand thinking they're as bad as Leverage, I think they've historically probably been a counterfactual negative for the movement, and don't have a good sense of whether things have improved. Thanks for clarifying. To be clear, I didn't say I thought they were as bad as Leverage. I said "I have less trust in CEA's epistemics to necessarily be that much better than Leverage's , though I'm uncertain here" 2Dancer22d I thought CEA started [https://forum.effectivealtruism.org/posts/9a7xMXoSiQs3EYPA2/the-history-of-the-term-effective-altruism] the movement? As I understood it, CEA was originally just a legal entity to save 80k and GWWC from having to both individually get charitable status, though GWWC had been around in some form since maybe 2007ish, and 80k for a year or two (and Givewell, which had started about the same time as CEA and arguably has as good a claim to having started it had no formal association with any of these orgs). The emerging movement might have taken its name from the new org, or maybe just started using the phrase in response to the poll result. At some stage IIRC, CEA started taking on more responsibilities and distanced itself, and eventually split from its child orgs. From that point on, I feel like they have generally not been well run - the staff seem to have been hired for enthusiasm and allegiance to the cause, and sometimes apparent nepotism (they seem to have hired internally for quite a few positions) rather than competence. As far as I can tell, staff have neither a carrot to motivate them or a stick: I know of only two examples of CEA employees being pushed out, one of who was CEO, and those were, as I understand it, for behaviour that was unambiguously termination-worthy (CEA may not want ... But good on you for being brave enough to publicly criticise your funding sources ("I have received EA funding in multiple capacities, and feel quite constrained in my ability to criticise CEA publicly") or people you like ("I like everyone I've interacted with from CEA"). 1RobBensinger22d I really like this comment, and I agree with it. 8Will Bradshaw23d ++++ 2BrownHairedEevee22d Why would being dismissed from CEA and being part of Leverage mean he has an axe to grind regarding SBF? Regarding 'top EA leaders' knowing about it (see further in the thread). 2Kerry_Vaughan22d If you'd like to investigate whether Leverage was a cult, there are now several additional sources of information available. One source is Cathleen's post [https://cathleensdiscoveries.com/LivingLifeWell/in-defense-of-attempting-hard-things] which is detailed, extensive, and written directly by a former employee. A board member conducted their own investigation [https://leverageresearch.org/inquiry-report-april-2022] into what Leverage could have done better between 2012 and 2019 by conducting interviews with former members of Leverage staff. You can also view Leverage's website [https://www.leverageresearch.org/] to learn more about what we've been working on post-2019. The fact that I work at Leverage is best explained by my having a very different view of the organization's history and current work than you do. In any case, I don't see why disagreements about the value of Leverage's current or past work have anything to do with the specific claims I've made about what happened at Alameda in 2018. I’d also recommend reading Zoe Curzi’s essay about her own (traumatic) experience at Leverage, the publishing of which was publicly supported by Leverage founder Geoff Anders. -8Guy Raveh22d I heard the same claim, from a different source: that SBF did something unethical at Alameda Research prior to founding FTX, that some EAs had left Alameda saying that SBF was unethical and no one should work with him, and that there were privately circulated warnings to this effect. (The person I heard this from hasn't spoken publicly about it yet as far as I know. They are someone with no previous or current involvement with FTX or Alameda Research, who I think is reporting honestly and is well positioned to have heard such things.) (EDIT: others along the rumor-path via which I heard this have now spoken on this thread, in greater detail than I have; so this comment is a duplicate report and should not be coutned.) + 1 for way more investigations and background checks for major donations, megaprojects, and association with EA. 6DaneelO22d I think this suggests that the EA orgs which had close ties to FTX and SBF should have investigations performed by outside parties. If this is true it makes the situation even worse than it appears at the moment since it could have been prevented by having higher ethical standards. -15Guy Raveh22d Thank you so much for your time, dedication, and efforts. It seems like, for many of us, difficult times lay ahead. Let us not forget the power of our community - a community of brilliant, kind-hearted, caring people trying to do good better together This is a crisis - but we have the ability to overcome it. I was really looking forward to maybe implementing impact markets in collaboration with Future Fund plus FTX proper if you and they wanted, and feel numb with regard to this shocking turn. I really believed FTX had some shot at 'being the best financial hub in the world', SBF 'becoming a trillionaire', and this longshot notion I had of impact certificates being integrated into the exchange, funding billions of dollars of EA causes through it in the best world. This felt so cool and far out to imagine. I woke up two days ago and this dream is now ash. I have spiritually entangled myself with this disaster. I don't want to be the first commenter to be that guy, and forgive me if I'm poking a wound, but when you have the time and slack can you please explain to us to what extent you guys grilled FTX leadership about the integrity of the sources of money they were giving you? Surely you had an inside view model of how risky this was if it blew up? If it's true SBF has had a history of acting unethically before (rumors, I don't know), isn't that something to have thoroughly questioned and spoken against? If there was anyone non-FTX who could have pressured them to act ethically, it would have been you. As an outsider it felt like y'all were in a highly trusted concerted relationship with each other going back a decade. In any case, thank you for what you've done. Sven Rone should've won a prize in the Red Teaming contest[1]: # The Effective Altruism movement is not above conflicts of interest [published Sep 1st 2022] # Summary Sam Bankman-Fried, founder of the cryptocurrency exchange FTX, is a major donator to the Effective Altruism ecosystem and has pledged to eventually donate his entire fortune to causes aligned with Effective Altruism. By relying heavily on ultra-wealthy individuals like Sam Bankman-Fried for funding, the Effective Altruim community is incentivized to accept political stances and moral judgments based on their alignment with the interests of its wealthy donators, instead of relying on a careful and rational examination of the quality and merits of these ideas. Yet, the Effective Altruism community does not appear to recognize that this creates potential conflicts with its stated mission of doing the most good by adhering to high standards of rationality and critical thought. In practice, Sam Bankman-Fried has enjoyed highly-favourable coverage from 80,000 Hours, an important actor in the Effective Altruism ecosystem. Given his donations to Effective Altruism, 80,000 Hours is, almost by definition, in a conflict of interest ... I wrote that comment from over a month ago. And I actually followed it up with a more scathing comment that got downvoted a lot, and that I deleted out of a bit of cowardice, I suppose. But here's the text: Consider this bit from the origin story of FTX In 2019, he took some of the profits from Alameda and8 million raised from a few smaller VC firms and launched FTX. He quickly sold a slice to Binance, the world’s biggest crypto exchange by volume, for about $70 million. Binance, you say? This Binance During this period, Binance processed transactions totalling at least$2.35 billion stemming from hacks, investment frauds and illegal drug sales, Reuters calculated from an examination of court records, statements by law enforcement and blockchain data, compiled for the news agency by two blockchain analysis firms. Two industry experts reviewed the calculation and agreed with the estimate.

Separately, crypto researcher Chainalysis, hired by U.S. government agencies to track illegal flows, concluded in a 2020 report that Binance received criminal funds totalling $770 million in 2019 alone, more than any other crypto exchange. Binance CEO Changpen ... Wow, I didn't see it at the time but this was really well written and documented. I'm sorry it got downvoted so much and think that reflects quite poorly on Forum voting norms and epistemics. Moreover, Sven Rone is a pseudonym. The author used a pen name astheir views were unpopular and underappreciated at the time; they likely feared career repercussions if they went public with it. It's unfortunate that this was the environment they found themselves in. Seconded. This whole saga has really made me sour on some already mixed views on EA epistemics. I find myself having a mixed opinion of how EA responded. It wasn't outright terrible epistemics, unlike most of the world reacting to a similar event, but there were real failures of epistemics. On the other hand, there was also successes in EA epistemics, as well. I think the post ended up around 0 or 1 karma, is that right? (I mean before people changed their voting based on hindsight!) I think it's important to distinguish between "got downvoted a lot but ended up at neutral karma" vs. "got downvoted double digits into no longer being visible." The former reflects somewhat poorly on EA, the latter very poorly. I think the most informative signal here is not the exact karma that comment ended up with but rather that the author ended up deleting it despite believing that what he was saying was potentially important and not receiving any reasons to think he was wrong. A culture where people feel compelled to silence themselves is worse than one where some comments are wrongly downvoted without much consequence to the author. I think the most important data points here are any comments that were left, and the net karma of the comment. People have in fact been known to overreact, or react in idiosyncratic ways, in forum discussions; I haven't seen the thread in question, but if the responses were friendly and the comment got ~0 net karma, then that would be a large update for me. I definitely took "that got downvoted a lot" to mean that the comment got a lot of net downvotes, not just that people offset its upvotes to keep it around a neutral 0. I think it's pretty bad to describe vote patterns that misleadingly, if it was hovering around 0. 3Lukas_Gloor22d Good point. :S 2Sam Elder20d Are we talking about this deleted comment [https://forum.effectivealtruism.org/posts/YgbpxJmEdFhFGpqci/winners-of-the-ea-criticism-and-red-teaming-contest?commentId=fSZQdKbRMJxXMxX9j] ? It has 6 overall karma in 9 votes, and -3 agreement in 5 votes. 2Lukas_Gloor20d No, I was talking about Stuart Buck's initial comment in that same thread, which is still up and now has high upvotes. But Stuart also mentioned he deleted a second comment after it got downvoted too, so that must be the one you're linking to. (We also don't know if some people retroactively upvoted the deleted comment, it's at +6 now but could've been negative at the time of deletion. I think I'm still able to vote on the deleted comment – though maybe that's just because I had already voted on it before it got deleted [strong upvote and weak agree vote]). 3Sam Elder20d Either way it seems highly unlikely that the deleted comment I linked to had lots of negative votes. It had a few disagree votes but very likely not more than 1-2 karma downvotes. I like how Hacker News hides comment scores. Seems to me that seeing a comment's score before reading it makes it harder to form an independent impression. I fairly frequently find myself thinking something like: "this comment seems fine/interesting and yet it's got a bunch of downvotes; the downvoters must know something I don't, so I shouldn't upvote". If others also reason this way, the net effect is herd behavior? What if I only saw a comment's score after voting/opting not to vote? Maybe quadratic voting could help, by encouraging everyone to focus their voting on self-perceived areas of expertise? Commenters should be trying to impress a narrow & sophisticated audience instead of a broad & shallow one? EDIT: Another thought: If there was a way I could see my recent votes, I could go back and reflect on them to ensure I'm voting in a consistent manner across threads I think that what FTX is accused of this comment is legitimately way more something where a charitable recipient is not morally obliged to demand this level of careful checking of everything, because our civilization is just not actually able to support this level of competency pornography. Stealing your customers' funds is a very different matter from "some of the people who use our services are criminals". Why, MIRI has in the past accepted matching funds from Google, which I'm sure profits a whole lot off criminals using their services! And some of those criminals may even be bad people! But you can't, actually, run a post-agricultural civilization on the principle of everybody who engages in every transaction checking out the full moral character of everybody who transacts with them. If you did try to build clever infrastructure for that, its first use on the margin would be by the right to hunt down sex workers (as already occurs with Visa) and by the left to hunt down people who said bad things on Twitter. In a hunter-gatherer tribe it maybe makes sense to demand that people not transact with that bad guy over there; it scales as far as it needs to scale. ... IMO the amount of diligence someone ought to perform on their counterparties' character is different in different circumstances. "This person is one of hundreds of people I transact with every week" carries different obligations than "This person is one of the four big donors who fund my organization" carries different obligations than "This person has been my only source of income for the past two years". Different EAs were at different points along this spectrum. 5Stuart Buck21d I generally agree with you, but in this case SBF 1) hired a high-level person with a long history of fraud (you don't see Asana or Stripe doing this); and 2) described his own business as a Ponzi scheme (see https://www.bloomberg.com/news/articles/2022-04-25/sam-bankman-fried-described-yield-farming-and-left-matt-levine-stunned [https://www.bloomberg.com/news/articles/2022-04-25/sam-bankman-fried-described-yield-farming-and-left-matt-levine-stunned] ). It was obvious that he was up to no good. "But Sequoia" -- I'm not convinced that they did any due diligence, judging by what they published on their own website: https://twitter.com/zebulgar/status/1590394857474109441 [https://twitter.com/zebulgar/status/1590394857474109441] It's not the only occasion when it looks to me like "top" VC firms leapt into investments out of FOMO, with zero effort at due diligence: https://medium.com/swlh/why-are-investors-eager-to-lose-money-on-health-tech-f8c678ccc417 [https://medium.com/swlh/why-are-investors-eager-to-lose-money-on-health-tech-f8c678ccc417] Quoting from the article you linked about the involvement of Daniel Friedberg, FTX's Chief Regulatory Officer, in a previous scandal: In 2008, online poker site Ultimate Bet (UB) publicly confirmed rumors that certain individuals had utilized a little-known feature of the site’s software to view players’ hole cards during hands. This so-called ‘god mode’ allowed a number of ‘super users’ to cheat opponents out of tens of millions in poker winnings. The site’s operators begrudgingly paid out a few million to the loudest complainers and folded the site’s operations into a sister site (which was dealing with its own scandals). In 2013, an audio recording surfaced that made mincemeat of UB’s original version of events. The recording of an early 2008 meeting with the principal cheater (Russ Hamilton) features Daniel S. Friedberg actively conspiring with the other principals in attendance to (a) publicly obfuscate the source of the cheating, (b) minimize the amount of restitution made to players, and (c) force shareholders to shoulder most of the bill. On the tape, Daniel S. Friedberg tells Hamilton that he doesn’t want news of the cheating scandal to get out, but if it must, the “ideal thi ... Sorry this text got heavily downvoted? If so, we should be ashamed. "They tell you to do your thing but they don't mean it. They don't want you to do your own thing, not unless it happens to be their thing, too. It's a laugh, Goober, a fake. Don't disturb the universe, Goober, no matter what the posters say.” - Robert Cormier, the Chocolate War Yes, we should. People hesitate or are averse to bringing issues up with authorities/communities due to fears of being punished. As groups collectivize and become increasingly memetically homogeneous, that which coincides with the solidification of power/influence/financial structures and hierarchies, dissent of any form becomes decreasingly tolerated. It becomes safer/easier to criticize EA as an outsider than as member who simultaneously want to grow in EA, be well received by potential EA organization employers, and rise up the oft unstated hierarchies that developed as EA blossomed. Until this debacle, SBF was lionized beyond comparison by the major community organizations. And moreover, he was closely associated with EA giants via the foundation/future fund and other projects. He had excellent PR presence due to the constant EA affiliated media attention. He was 80k's paragon of earning to give... 3Denkenberger20d By the way, it looks like the comment is now heavily upvoted. I've seen this happen quite a few times, so it seems like it might be good to withhold judgment about the net votes for a day or two. But of course it could be that it became highly upvoted because of reactions like this, so I'm not sure what the best course of action is. 9Sharmake23d I don't follow crypto, or it's space, but this seems like a bad habit or norm to downvote pieces that criticize EA's questionable relationship to crypto. Cryptocurrency doesn't actually work, and only is there for scams and fraud. Not surprising that FTX collapsed. I think you may be getting a lot of disagree-votes because I don't think crypto was the issue here. People who just have USD sitting in FTX right now lost their money too. FTX shouldn't have been risky. It wasn't a DAO, or based entirely off some token or chain, it was an exchange. It should have just been connecting people who wanted to buy crypto with people who wanted to sell crypto, and taking a fee for doing this. The exchange itself shouldn't be taking any risk. The reason as to how looks at least in part to do with leveraged transactions, allowing customers to buy more crypto by supplementing their purchase with a loan. But we've let leveraged transactions happen with stock for a hundred years. This looks a lot more like garden-variety financial crime than some problem with crypto. Here's a quote from former US Treasury Secretary Larry Summers in a recent Bloomberg interview that backs up some of the claims in this comment: A lot of people have compared this to Lehman. I would compare it to Enron - the smartest guys in the room, not just financial error, but certainly from the reports, whiffs of fraud. Stadium namings very early in a company's history. Vast explosion of wealth that nobody quite understands where it comes from. [...] I think this is probably less about the complexities of the nuances of the rules of crypto regulation and more about some very basic financial principles that go back to financial scandals that took place in ancient Rome. Sorry for misfiring here, I'll retract my comment. [This comment is no longer endorsed by its author]Reply 4Jason22d The relation to crypto is that the bulk of crypto is poorly regulated. Some of that is solvable -- well regulated exchanges should be possible. The extreme volatility also increases the temptation toward fraud. So the fraud risk is higher than in a well-regulated industry. I'd submit that a well-regulated and managed exchange is going to find it much harder to achieve a stratospheric valuation, and other parts of crypto are harder to regulate well. So some skepticism toward huge crypto-linked donors is warranted. 3Geoffrey Miller22d More crypto regulation is coming, and many crypto protocols have worked hard already to be regulatory-compliant. But regulation won't be uniform across jurisdictions; there will always be loopholes that allow regulatory arbitrage. Some exchanges, such as Coinbase and Kraken, are based and regulated within the US, and are subject to much stricter oversight than FTX -- which seems to have been deliberated based in Hong Kong and then the Bahamas precisely in order to avoid US regulatory oversight. (Arguably, this should have been a red flag in terms of EA's relationship with FTX). The US, UK, or EU can regulate all they want, but crypto finance is a global business, and there are plenty of less-regulated havens willing to host crypto businesses. Hopefully crypto investors, traders, and users will become savvier about checking where businesses are operating, and what regulatory scrutiny they're subject to. 1Jason22d Agreed on that. My point was that it would be a lot harder for an individual to get super-rich quick in a regulated market. No sane regulator is going to allow a regulated party to risk customer assets for the party's benefit, and few will allow crazy leverage. And the whole thing will require significantly more of a buffer in fiat currency, again limiting any single person's ability to get megarich. In short, I think there are few ways for a well-regulated exchange to be stratospherically profitable. So people should not expect the rise of new crypto megadonors who hail from regulated backgrounds. 2Zach Furman22d I would agree with this. Separate from the object-level causes of the current crisis, crypto as an industry has accepted and normalized a lack of accountability that other industries haven't. And I agree that lack of regulation and high volatility make fraud more likely. I would want to avoid purely focusing on crypto, because I think the meta-lesson I might take away is less "crypto bad" and more "make sure donors and influential community members are accountable," whether that be to regulators, independent audits, or otherwise. (And accountable in a real due diligence sense, because it's easy for that word to just be an applause light.) But yes, skepticism of crypto-linked donors would be justified under this framework. 2Sabs22d I have no idea why this comment is no longer endorsed by its author because it's entirely correct. Not only is crypto a great way to scam people because transactions can't be reversed & there's virtually no regulation for most of the space, the fact that it's so hard to make money in crypto across an entire cycle means that entities have a huge incentive to resort to scamming. I can tell you why I downvoted it. Cryptocurrency doesn't actually work False, it works just fine. It's a token that can't be duplicated and people can send to each other without any centralized authority. and only is there for scams and fraud. There are indeed a lot of those, but scams and fraud were very clearly not the intention of its creators. Realistically they were cryptography nerds who wanted to make something cool, or libertarians with overly-idealistic visions of the future. Not surprising that FTX collapsed. Clear hindsight bias. This person should have made some money betting against FTX before it collapsed and then I'd take them more seriously. Basically, the comment is just your standard "cryptocurrency bad" take, without any attempt at justifying their claims or even saying much of anything other than expressing in an inflammatory way that they don't like cryptocurrency. "This person should have made some money betting against FTX before it collapsed and then I'd take them more seriously." this is naive EMH fundamentalism not everything can be shorted, not everything can be shorted easily, not everything should be shorted, markets can be manipulated. Especially the crypto market. It both can be the case that people 100% think X is a fraud, and X collapses, and shorting X would have been a losing trade over most timeframes. "Never short" is an oversimplification but honestly not a bad one. 2Czynski20d Most of that isn't even clearly bad, and I find it hard to see good faith here. Your criticism of Binance amounts to "it's cryptocurrency". Everyone knows crypto can be used to facilitate money laundering; this was, for Bitcoin, basically the whole point. Similarly the criticism of Ponzi schemes; there were literally dozens of ICOs for things that were overtly labeled as Ponzis - Ponzicoin was one of the more successful ones, because it had a good name. Many people walked into this with eyes open; many others didn't, but they were warned, they just didn't heed the warnings. Should we also refuse to take money from anyone who bets against r/wallstreetbets and Robinhood? Casinos? Anyone who runs a platform for sports bets? Prediction markets? Your logic would condemn them all. FTX would prefer that the crypto sector stay healthy, and backstopping companies whose schemes were failing serves that goal. That is an entirely sufficient explanation and one with no clear ethical issues or moral hazard. Even in retrospect, I think this was bad criticism and it was correct to downvote it. 9Stuart Buck20d My criticism of Binance was not "it's cryptocurrency." My criticism of Binance was that at the very time that that SBF allied with Binance, it was a "hub for hackers, fraudsters and drug traffickers." Apparently your defense of SBF is that "everyone knows" crypto is good for little else . . . but perhap if someone enters a field that is mostly or entirely occupied by criminal activity, that isn't actually an excuse? As for backstopping other scams and frauds, that isn't a way to make sure that the "crypto sector stays healthy" (barring very unusual definitions of the word "healthy"), and in actuality, we're now seeing evidence that FTX was just trying to extract assets from other companies in a desperate attempt to shore up their own malfeasance and fraud. https://twitter.com/AutismCapital/status/1591569275642589184 [https://twitter.com/AutismCapital/status/1591569275642589184] -1Czynski16d Yeah, still not seeing much good faith. You're still ahead of AutismCapital, though, which is 100% bad faith 100% of the time. If you believe a word it says I have a bridge to sell you. 3Stuart Buck14d Is this Sam in disguise? You're literally the only person in existence who seems to think it was somehow unfair to be suspicious (and correctly so!) of SBF for having hired a chief compliance officer with a long history of fraud, and of his pattern of trying to buy up other people's frauds/scams. 6Stuart Buck20d The only flaw in my earlier comment is that I was too charitable towards SBF in suggesting that there might be some plausible excuse for the multiple red flags I noticed. Thanks for this! I echo Lizka's comment about linkposting. In light of the recent events I'm struggling a bit with taking my hindsight-bias shades off, and while I scored it reasonably highly, I don't think I can fairly engage with whether it should have received a prize over other entries even if I had the capacity to (let alone speak for other panelists). I do remember including it in the comment mainly because I thought it was a risk that didn't receive enough attention and was worth highlighting (though I have a pretty limited understanding of the crypto space and ~0 clue that things would happen in the way they did). I think it's worth noting that there has been at least one other post on the forum that engaged with this specifically, but unfortunately didn't receive much attention. (Edit: another one here) Ultimately though, I think it's more important to think about what actionable and constructive steps the EA community can take going forward. I think there are a lot of unanswered questions wrt accountability from EA leaders in terms of due diligence, what was known or could have been suspected prior to Nov 9th this year, and what systems or checks/balances were in place... 6Random23d It seems it was entered, according to the (second) comment from Bruce here: Winners Red Teaming [https://forum.effectivealtruism.org/posts/YgbpxJmEdFhFGpqci/winners-of-the-ea-criticism-and-red-teaming-contest ] 4Greg_Colbourn23d Thanks (link to the comment [https://forum.effectivealtruism.org/posts/YgbpxJmEdFhFGpqci/winners-of-the-ea-criticism-and-red-teaming-contest?commentId=JjvpwGKeaj4w5FTzP] ). I think those entries really should've been put on the EA Forum as posts to be interacted with (like with the Future Fund AI Worldview Prize[1] [#fnliaheitd3g9]) 1. ^ [#fnrefliaheitd3g9]Which I imagine is no longer happening :( 8Lizka23d Yeah, I can confirm that we evaluated that submission. Re: putting them on the Forum — we didn't have the capacity to do that (and I'm not sure it would have been helpful to do that for all the submissions), but in general, I really encourage people to link-post relevant content to the EA Forum [https://forum.effectivealtruism.org/posts/8yDsenRQhNF4HEDwu/link-posting-is-an-act-of-community-service] . So, you could link-post this (or similar content in the future). [I should note that I have low capacity right now and might not reply to this thread. Apologies in advance!] 4Czynski20d Strongly disagree. That criticism is mostly orthogonal to the actual problems that surfaced. Conflicts of interest were not the problem here. 2Greg_Colbourn20d I'd regard incentive to discount highly immoral business practices (e.g. what happened with Alameda in 2018) as stemming from a conflict of interest (i.e. interest 1: promote integrity in EA; interest 2: get lots of money from SBF for EA. These were in conflict!) 1Czynski16d Again, that's orthogonal to the actual problems that surfaced. 7Greg_Colbourn16d I wouldn't say orthogonal, more upstream. If SBF had been shunned from the community in 2018, would we be in this situation now? Sure, he might still have committed massive fraud with the ends of gaining wealth and influence, but the focus would be on the Democrats, or whatever other group became his main affiliation. 6Czynski5d No, you're thinking about it entirely wrong. If everyone who did something analogous to Alameda 2018 was shunned, there probably wouldn't be any billionaire EA donors at all. It was probably worse than most startups, but not remarkably worse. It was definitely not a reliable indicator that a fraud or scandal was coming down the road. 6Greg_Colbourn5d Dustin Moskovitz and Jaan Tallinn were already EA ~billionaire donors well before 2018. They haven't done anything analogous to what SBF/FTX/Alameda did. What examples are you thinking of? FTX had received several billion dollars in funding from major investors. One was a province pension fund, so it wasn't just crypto folks. That generally involves having the investors' accountants do substantial due diligence on the target firm's financials. That tells me that either the books were fairly clean at the time of investment or they were cooked in a way that even the due dilligence specialists didn't detect. It's not clear to me how the Future Fund people, who to my knowledge are not forensic accountants or crypto experts, would have had a better ability to pick up on funny business. So I don't see why it would be unreasonable for them to have relied on third-party expert vetting. From what I understand (please correct me if I'm wrong), FTX didn't have a CFO, it's COO was a friend with no experience, and it didn't have a proper board of directors. Clearly, that flimsy corporate governance would not pass a standard due diligence test. EDIT: This flow chart of shells nested in shells, like Russian dolls, speaks to why the company's governance should have been a red-flag. this flow chart of shells nested in shells, like Russian dolls, speaks to why the company's governance should have been a red-flag. I don't think a highly branched company structure is a red flag: my understanding is that to operate a financial business legally across many jurisdictions you generally need to have subsidiaries in each jurisdiction. Ex: https://wise.com/help/articles/2974131/what-are-the-wise-group-entities In the autopsy, the biggest red flag will probably be the lack of appropriate internal controls. One should not be able to move that kind of money without vetting by staff with appropriate background and independence, but no ownership interest. Based on the reported en masse resignation of the bulk of legal and compliance staff, it seems that it was technically possible to transfer billions in customer assets to the CEO's company without legal/compliance involvement. I think the class of issues that would make it inappropriate to accept donations is much narrower than the issues that would and should make a public investor (like a province pension fund) decline to invest. Few private businesses are going to let an outsider come in on a regular basis, conduct a hard look at sensitive internal documents, and potentially publish derogatory information to the public. Even for investors, this kind of stuff is generally done under a heavy NDA and for good reason. That would make it extremely difficult to do this on a regular basis -- so any scrutiny would at best catch fraud that existed at the time of scrutiny. I wouldn't be very confident in the level of due diligence undertaken by supposedly sophisticated investors: https://twitter.com/zebulgar/status/1590394857474109441 This just isn't plausible on reasonable priors. You need to assume that multiple investment firms working in different sectors, whose survival in a highly competitive environment in large part depends on being skilled at scrutinizing a company's financials, would miss warning signs that should have been apparent to folks with no relevant domain expertise. See also Eliezer's Twitter thread. ETA: Alexander: Some people are asking whether people who accepted FTX money should have “seen the red flags” or “done more due diligence”. Sometimes this is from outsider critics of effective altruism. More often it’s been effective altruists themselves, obsessively beating themselves up over dumb things like “I met an FTX employee once and he seemed to be frowning, why didn’t I realize that this meant they were all committing fraud?!” Listen: there’s a word for the activity of figuring out which financial entities are better or worse at their business than everyone else thinks, maximizing your exposure to the good ones, and minimizing your exposure to the bad ones. That word is “finance”. If you think you’re better at it than all the VCs, billionaires, and traders who trusted FTX - and better tha ... 3DaneelO21d I would disagree, there are numerous examples such as Theranos and WeWork which show that sophisticated investors do not necessarily scrutinize potential investments thoroughly. Thus I don't think assuming they do is a good prior. I think this is actually a reason these problems happen, since everyone else assumes that Respectable Company/Person X has scrutinized it. I am making a comparative, not an absolute, claim: however bad the professionals may be, it is unreasonable to expect outsiders to do better. 2DaneelO21d I agree with the point that in general one should expect less from "unsophisticated" investors/parties than from sophisticated ones. I do not disagree with that. I was disagreeing with "This just isn't plausible on reasonable priors." which seemed to mean that you disagreed with Stuart's comment. But I also don't think VC scrutiny is necessarily a high bar in general in the absolute sense, and Stuart has posted some warning signs here in other comments such as the hiring of Friedberg. Then considering how important FTX and SBF was to the EA community it could have been investigated more, i.e. the low VC scrutiny bar could have been surpassed by hiring experts or something similar. To a VC firm this is just another losing bet among many they expect to make. This is why I don't think the comparison with VC firms is very apt. 9Pablo21d Stuart's comment was in reply to the claim that "It's not clear to me how the Future Fund people, who to my knowledge are not forensic accountants or crypto experts, would have had a better ability to pick up on funny business." I disagreed with Stuart's comment in the sense that I disputed the reasonableness of expecting unsophisticated outsiders to do better because sophisticated investors sometimes perform poorly. I did not mean to dispute that sophisticated investors sometimes perform poorly; indeed, there's plenty of evidence of that, including the evidence you provide in your comment. 4DaneelO21d Yeah that makes sense, I think I overinterpreted your comments. 7Pablo21d In retrospect, I think my original comment was insufficiently clear. Anyway, thanks for the dialogue. 5Pat Myron21d And that's what Sequoia proudly publicly posted themselves 5Sharmake23d Or the best auditors are inadequate, and overlooked fairly obvious flaws for some reason. can you please explain to us to what extent you guys grilled FTX leadership about the integrity of the sources of money they were giving you? Surely you had an inside view model of how risky this was if it blew up? This sounds a bit hindsight-bias-y to me; we know to poke at this specific topic now because we know what happened. SBF claims to not have known himself that this was happening, which I take to mean that either this info was super siloed or buried somehow, or that Sam is lying. (And is relying on few-to-no people knowing the truth, or someone would immediately call him out on the lie.) The idea that SBF didn't know what was happening is farcical. You don't unknowingly loan out$10bn of customer funds, which you then lose on bad bets, and then try and cover up your insolvency. I think it's healthy to wait for a clearer picture of what happened before making any summary judgement, but we know enough to say that SBF was not an honest actor.

To be honest, I'm at a point now where I'm putting significant weight on lying. Some evidence here that FTX bailed out Alameda for ~$4B in FTT on Sep 28th. There are the blockchain transactions (disclaimed by SBF at the time), and the resignation of a high-profile figure (President of FTX.US) the day before. (Note that whilst this doesn't look good, it's still inconclusive. I'm sure the truth will come out eventually.) I agree that there's a lot of hindsight bias here, but I don't think that tweet tells us much. My question for Dony is: what questions could we have asked FTX that would have helped? I'm pretty sure I wouldn't have detected any problems by grilling FTX. Maybe I'd have gotten some suspicions by grilling people who'd previously worked with SBF, but I can't think of what would have prompted me to do that. 9Lukas_Gloor22d There were IMO some orange flags (such as the connection to the questionable lawyer who also works with tether), but admittedly think it's difficult to notice such things when there's an aura of success around someone. I think it isn't just hindsight, though. I think people need to get a lot better at being cynical, because it's important. For instance, it was odd how FTX positioned itself as the savior of crypto by proposing to buy out entities like Voyager and Blockfi and then it comes out that Alameda owes them money. They said they could "pay anytime," but it still looked weird. Please feel free to "be that guy" as hard as possible when we are talking about massive financial fraud. Hope you're feeling okay Dony. [on phone] Thank you so much for all of your hard work managing the fund. I really appreciated it and I think that it did a lot of good. I doubt that you could have ever have reasonably expected this outcome so I don't hold you responsible for it. Reading this announcement was surprisingly emotional for me. It made me realise how many exceptionally good people who I really admire are going to be deeply impacted by all of this. That's really sad in addition to all the other stuff to be sad about. I probably don't have much to offer other than my thoughts and sympathy but please let me know if I can help. I suppose that I should disclose that I recently received a regrant from FTX which I will abstain from drawing on for the moment. I don't think that this has much, if any, relevance to my sentiments however. I would not draw on that grant for quite some time, if ever: you should be worried about clawbacks. 9Leon_Lang23d I have no idea under what circumstances clawbacks can happen. If you have good reasons to believe this is plausible, then it seems worth it to write a top level post on it. The comment below that this is like Bernie Madoff is not right as far as I can see. This is a different situation, with different facts - including that we have, as yet, no idea what those facts are! Your situation will also be individual - if you took the funds as a limited company is different to if you took them individually, for example, with different effects most likely. It is also entirely unknown what is happening. Nothing has been made clear officially, no one knows what's going on and you - importantly - had nothing to do with any of that stuff that is being potentially alleged (not yet actually alleged by any authority). I'm not giving legal advice here. I'm just stating that being calm is the right response and that googling Bernie Madoff (as suggested below), won't most likely be of any help. -47Sabs23d 0[comment deleted]23d Hey team -- thank you for all the work you did. The Future Fund has been tremendously inspiring to see. I'll reach out to you about how we (myself or Protocol Labs) might be able to help. 3Dawn Drescher23d Pooling the expertise of the Future Fund team and Protocol Labs would be amazing! <3 I commend you on your moral leadership and I join everyone else in the comments in expressing gratitude for the tremendous good you've done so far. However, I'm curious about your decision to resign. I get the moral justification, but surely there are many grantees with many questions who'd be able to get better answers were you still within Future Fund. Something as simple as access to documents or previous emails would enable you to better support grantees who are likely in significant distress. Why did you see it as imperative to resign effective immediately? Why not at the very least see out your notice period? 9sphor23d How does it take moral leadership to distance yourself from and condemn massive fraud? Even entirely selfish actors would do the same. 7Markus Amalthea Magnuson23d I'm curious about this as well. Does leaving immediately not impede the chances of getting a better (I'd never dare say "full") picture of what went down? Additionally, in terms of accountability, I guess now we'll never know or have records of (from emails etc.) who knew what and when. 5Jason23d I don't think staying on would add to what the insolvency trustee, regulatory authorities, and likely criminal prosecutors will uncover. The court has already appointed a liquidation trustee whose mission is preserving assets and does not include working with EA. Its unclear to me whether the trustee is in control of the FTX Foundation now, but the statement did say related entities. The FTX principals are doubtless preoccupied and are presumably attuned enough to legal exposure to not be having unnecessary conversations. Thanks very much for posting this update! My main question re the Future Fund at the moment is: why does it seem like there weren't any ring-fenced funds under legal ownership by the Future Fund or the FTX Foundation? Are there any? Were there any when it was founded last year (i.e. presumably when FTX/Alameda was still solvent)? If not, why not? Did this not raise suspicions amongst any of you? I can imagine maybe SBF saying something like the max-EV thing to do is keeping all the funds in the for-profit companies to maximise their growth, and you going along with it because you trusted him (or you just independently agreed and didn't put any significant weight on FTX/Alameda collapsing or even just becoming less rich). Obviously an error in hindsight. Or maybe you kept asking about getting (more) ring-fenced funds, and kept getting fobbed off? That should've raised alarm bells if so! Sorry if this is a bit ranty and speculative, or too soon, or too accusatory, but I'm grasping for answers here. I'm grateful for everything you've done for the world and EA in your careers, but can't help feeling that you might've messed up a bit here. 5Milan_Griffes21d I asked some further questions in this direction here [https://forum.effectivealtruism.org/posts/WdeiPrwgqW2wHAxgT/a-personal-statement-on-ftx?commentId=3ZNGqJEpQSrDuRpSu] . 3Dancer19d https://www.nytimes.com/2022/11/13/business/ftx-effective-altruism.html 3Greg_Colbourn19d Thanks for linking. That should've raised alarm-bells, in hindsight. Could he not at least have donated illiquid assets to the Foundation, for them to liquidate as they see fit (and put the Foundation under independent control)? Although guess that still might not've helped much in this case with FTT and FTX stock collapsing. 3Aleks_K18d I think this (the fact that there is no endowment) was (or at least should have been) pretty well-known in the EA community from the point in time that the FTX Future Fund started to pay grants, as these came from all kinds of sources, but not from an endowed foundation. And it obviously would have been known to the people working for FTX Foundation from when they started working there. (And I would guess one reason that it didn't raise more alarm bells for lots of people in the EA community that learned about this, is probably that they put high trust in the people working for FTX Foundation.) 2ZekeFaux18d How much did the Future Fund actually pay out? The website lists$160 million in committed grants.
1Aleks_K18d
I agree this would be very useful information. In theory, the FTX Future Fund team should know this information but they probably are not allowed to share it. Of course, someone could try to collect this information by contacting all named FTX Future Fund grantees and it might be worth the effort to try to do this. (Though it's unclear who might be best suited to do that, given that they'd have to be trusted enough by all grantees for them to share their individual details with them.) Maybe the largest recipients (I think these are CEA and Longview) could start by stating how much they received.
1Greg_Colbourn18d
Made this into a post: Why didn't the FTX Foundation secure its bag? [https://forum.effectivealtruism.org/posts/YAsJCsvi4tBpCwuFB/why-didn-t-the-ftx-foundation-secure-its-bag]
2Greg_Colbourn18d
Can anyone find the original source for the "interview last month"? Clicking that link from the link above takes me to https://www.nytimes.com/2022/10/08/business/effective-altruism-elon-musk.html [https://www.nytimes.com/2022/10/08/business/effective-altruism-elon-musk.html] (a [https://archive.ph/qkejy]) which doesn't contain the quote.

Please let us know if there is anything we at GoodX can do to help. Our main project is to build an impact marketplace, but ultimately we want to get resources to where they are needed (as efficiently as possible).

(E.g., it wouldn’t be my first time running an emergency fundraiser to bail out customers of a failed venture.)

Strikes me as…premature? We’ll have a lot more clarity in the coming days, and resigning + questioning the ethics at FTX when we still fundamentally don’t know what happened doesn’t seem particularly productive.

If FTX just took risks and lost, this will look very dumb in hindsight. And if there turn out to be lots of unethical calls, we’ll have more than enough time to criticize them all to our hearts’ content. But at least we’ll have the facts.

Looking dumb is an acceptable risk. If the team prematurely resigned and there is still usable money . . . the usable money is presumably locked in the FTX Foundation and in DAFs, it is not lost.

Premature send, ETA: As far as "questioning the ethics at FTX," it would be very easy for FTX to have denied raiding customer funds if they didn't do it as reported. It's appropriate to draw the obvious inference that they did, and that alone is more than enough to "question[] the ethics at FTX" which is a pretty mild response to the news in my book.

The PR attention is at its height this week, the risk of "looking dumb" (which I think is very unlikely) is outweighed by the need to engage in damage control. No one will be listening if EA waits a few weeks to start distancing itself....

From The Snowball, dealing with Warren Buffett's son's stint as a director and PR person for ADM:

The second the FBI agents left, Howie called his father, flailing, saying, I don't know what to do, I don't have the facts, how do I know if these allegations are true? My name is on every press release. How can I be the spokesman for the company worldwide? What should I do, should I resign?

Buffett refrained from the obvious response, which was that, of his three children, only Howie could have wound up with an FBI agent in his living room after taking his first job in the corporate world. He listened to the story non-judgmentally and told Howie that it was his decision whether to stay at ADM. He gave only one piece of advice: Howie had to decide within the next twenty-four hours. If you stay in longer than that, he said, you'll become one of them. No matter what happens, it will be too late to get out.

That clarified things. Howie now realized that waiting was not a way to get more information to help him decide, it was making the decision to stay. He had to look at his options and understand as of right now what they meant.

If he resigned and they were innocent, he would lose friends an

...
6t3tsubo23d
The facts are plenty clear (with respect to the type of criminal activity taking place, if not the specifics or quantum) if you do some digging on twitter. Crypto-forensics have been having a field day and SBF himself has surprisingly been continuing to dig his grave deeper.

I would highly, highly recommend that people just wait up to 72 hours for more information, rather than digging through Twitter or Reddit threads.

Edit: This is not to imply that I have secret information - just that this is unfolding very quickly and I expect to learn a lot more in the coming days.

6Geuss23d
Why? Coin Desk's leak - which set off the death spiral - is clear enough. Multiple investors that SBF tried to get bail-out funds from have told the WSJ and FT that SBF admitted to loaning out customer funds to Alameda. Binance pulled out of the deal for a reason. There is plenty of data online about FTX's movements on the blockchain. And, of course, there's the obvious fact that SBF is now very publicly looking for \$8bn of funding to cover FTX's liabilities.
5Nathan Young23d
Feels like you are implying you have secret info, but it just seems extremely unlikely to me that this was anything other than huge mismanagement of customer funds against their wishes. What odds are you willing to bet that we will see it differently in 72 hours?

I don't think the bet suggestions (not just from you - there were a bunch in others' comments on your own post) are helping make the situation any less tense.

Edit: I also think the interpretation of "implying to have secret information" rather than "trying to de-escalate" is not really grounded, and results in your comment being combative in my eyes.

I think bets with real stakes can be a good de-escalation procedure! It's easy to fire increasingly heated claims back and forth while there's no concrete consequences, but when there's money on the line you have to back off and figure out what you actually believe, and then also once the bet is made there is less incentive to keep arguing while you wait for resolution.

3RAB23d
Didn’t mean to imply secret info, edited the comment above. That said, seeing most of their legal and compliance teams quit gives me much more serious reservations about illegal or unethical behavior. Edit: I think I retract this second part - I don’t know if everyone’s quitting now that they can’t pay salaries, or just the legal/compliance teams.

I've made this into a post on the forum, because I'm afraid it'll get buried in the comments here. Please comment on the forum post instead.

https://forum.effectivealtruism.org/posts/9YodZj6J6iv3xua4f/another-ftx-post-suggestions-for-change

I suggested that we would have trouble with FTX and funding around 6 months ago.

SBF has been giving lots of money to EA. He admits it's a massively speculative bubble. Crypto crash hurts the most vulnerable, because poor uneducated people put lots of money into it (Krugman). Crypto is currently small, but should be regulated and has potential contagion effects (BIS). EA as a whole is getting loose with it's money due to large crypto flows (MacAskill). An inevitable crypto crash leads to either a) bad optics leading to less interest in EA or b) lots of dead projects.

It was quite obvious that this would happen--although the specific details with Alameda were not obvious. Stuart Buck was the only one who took me seriously at the time.

# Below are some suggestions for change.

1. The new button of "support" is great, but I think EA forum should have a way to *sort* by controversiality. And, have the EA forum algorithm occasionally (some&...

9Sharmake22d
Being honest, I do genuinely think that climate change is less important than runaway AI, primarily because of both option value issues and the stakes of the problem. One is a big problem that could hurt or kill millions, while AI could kill billions. But I'm concerned that they couldn't simply state why they believe AI is more important than climate change rather than do this over-complicated scheme. Disagree, this would make transparency worse without providing much benefit. Disagree here because I don't want to see an EA forum that values controversial posts.
1SaraAzubuike22d
Hi, thanks for replying! I've made this into an EA forum post, instead because I'm afraid it'll get buried in the comments here. https://forum.effectivealtruism.org/posts/9YodZj6J6iv3xua4f/another-ftx-post-suggestions-for-change

Question just to double-check: are posts no longer going to be evaluated for the AI Worldview Prize? Given that is, that the FTX Future team has resigned.

I think it would be good if others stepped in to help see it through (perhaps offering smaller prizes), given how critical the answers are to determining EA resource allocation. Have asked Holden re OpenPhil fulfilling this role.

8Geuss23d
Why do you think it's any more important than the FTX Fund's other obligations? If there's to be a settlement matching partial assets to all of the fund's liabilities, it should done in an open and fair way. Maybe the assets are 0, in which case that becomes moot. My own view is that there are many other projects of equal or greater merit with funding commitments from the FTX Fund.
7Greg_Colbourn23d
That's reasonable. I guess from my perspective, I think the top EA grantmakers need persuading that p(doom|AGI) is significantly greater than 35%. If OpenPhil already think this, then that's great, but if they don't (and their probabilites are similar to the Future Fund's), then the Worldview prize is very important. Even if your probabilities are the same, or much lower, it's still very high Value of Information imo.

In the survey I did last year, four Open Phil staff respectively gave probability 0.5, 0.5, 0.35, and 0.06 to "the overall value of the future will be drastically less than it could have been, as a result of AI systems not doing/optimizing what the people deploying them wanted/intended".

That's just four people, and isn't necessarily representative of the rest of longtermist Open Phil, but it at least shows that "higher than 35%" isn't an unrepresented view there.

4Greg_Colbourn23d
Interesting, thanks. What about short timelines? (p(AGI by 2043) in Future Fund Worldview Prize terms)
7RobBensinger22d
Ajeya Cotra's median guess is that AGI is 18 years away [https://www.lesswrong.com/posts/AfH2oPHCApdKicM4m/two-year-update-on-my-personal-ai-timelines] ; the last time I talked to a MIRI person, their median guess was 14 years. So the Cotra and MIRI camps seem super close to me in timelines (though you can find plenty of individuals whose median year is not in the 2036-2040 range). If you look at (e.g.) animal welfare EAs vs. AI risk EAs, I expect a much larger gap in timeline beliefs.
2Jason23d
One could also argue for prioritizing funding for work that has already been done over work that has been approved but not yet done. If someone was going to receive a grant to do certain work and has it been pulled, that is unfair and a loss to them . . . but it's not as bad (or as damaging to the community / future incentives) as denying people payment for work they have already done. How this logic translates to a prize program is murky. But unless you believe that the prize's existence did not cause people to work more (i.e., that the prize program was completely ineffective), its cancellation would mean people are not going to be paid for work already performed. Of course, it might be possible to honor the commitment made for that work in some fashion that doesn't involve awarding full prizes.

Potential Help for FF Grantees.  I work at a major philanthropic organization, Stand Together, on technology and innovation related efforts.  I was a big fan of Future Fund's ambition and methods, even where I didn't share your priors.

At Stand Together, we work on a wide range of issues, all seeking to break the barriers that prevent individuals from reaching their true potential. On technology, we think technological innovation has been the primary driver of widespread human prosperity and we are looking to promote both a culture that embraces innovation rather than fears it and a regulatory environment that enables it.

If you are a Future Fund grantee interested in alternative funding and any of the above seems to line up with your work, please reach out: nchilson@standtogether.org.

And best of luck to everyone.

Wishing much strength to everyone affected by this. Let's support each other and get through this together.

as a non EA reading this thread, on balance, makes me really happy. You guys just have some good old fashioned cleansing to do and you'll be fine.

FWIW, everyone who's had any dealings with the Alameda crew knew that they were the worst kind of trash - we just thought that meant they have so much money that surely they don't need to steal ours.

cheers.

It seems like there are quite a lot of people/orgs who made plans based on promised money that now seems unlikely to arrive. Is there a lesson that can be learned about how to reduce risk in grant awarding e.g. by waiting until funds are securely in the foundation's hands? Or is there no way to avoid this risk given potential clawbacks, even in cases of bankruptcy that don't involve any fraud?

Thank you for your good work over the last months, and thank you for your commitment to integrity in these hard times. I'm sure this must also be hard for you on a personal level, so I hope you're able to find consolation in all the good that will be created from the projects you helped off the ground, and that you still find a home in the EA community.

I trust you guys to decide that this is the right time to resign, but I do hope as a community that we are able to hold value of our friendships together with the importance of holding people who made mistakes to account, without either one negating the other. We don't yet know what kind of ethical errors Sam made, but the larger those mistakes are, the more important it is that we offer friendship of a kind that is compatible with holding people to account.

In his post announcing the new found wealth of EA movement stemming from FTX Will included this argument for why charitable enterprises are more dangerous than for profit companies:

There’s one huge difference between aiming to do good and aiming to make profit. If you set up a company aiming to make money, generally the very worst that can happen is that you go bankrupt; there’s a legal system in place that prevents you from getting burdened by arbitrarily large debt. However, if you set up a project aiming to do good, the amount of harm that you can do is

...

This feels like a weird interpretation of Will's comment, which doesn't (in my view) imply that for-profit companies can't do a lot of harm, but rather that if you start a company with the sole goal of making a profit, usually the worst outcome (with regards to your goal of making a profit) is that you go bankrupt.

6Jonathan Paulson22d
As FTX just spectacularly demonstrated, Will was wrong. This is because even though FTX was ostensibly started with the sole goal of making a profit, it turns out there were other important implicit goals like “don’t steal billions of dollars from thousands of people”, implicit goals like that always exist, and failure to meet those implicit goals is very bad.
2Sharmake21d
This sounds like a human form of alignment failure, specifically, the What Failure Looks Like story part I. Here's a link to it: https://www.lesswrong.com/posts/HBxe6wdjxK239zajf/what-failure-looks-like [https://www.lesswrong.com/posts/HBxe6wdjxK239zajf/what-failure-looks-like]
5Sharmake22d
Should have called it, but I'll do it now: It's a double standard applied, so the comparison is not what you think.