All of Xavier_ORourke's Comments + Replies

Love to see it.

The true valuation is probably a lot more than $26b USD - there was recently a secondary sale which valued company at closer to 42 billion. If the company became public, and were to trade at a revenue multiple similar to Figma's (extremely high at like 30x right now) - this would set the company's valuation a lot higher than that

I might not be the target audience for this proposal (my EA involvement weakened before FTX, and I'm not on track for high-leverage positions) - so take this perspective with appropriate skepticism. I'm also making predictions about something involving complex dynamics so good chance I'm wrong...

But I see fundamental challenges to maintaining a healthy EA movement in the shape you're describing. If we really encourage people to be vocal about their views in the absence of strong pressure to toe a party line - we can expect a very large, painful, disruptive... (read more)

I think an important consideration being overlooked is how comptetntly a centralised project would actually be managed.

In one of your charts, you suggest worlds where there is a single project will make progress faster due to "speedup from compute almagamation". This is not necessarily true. It's very possible that different teams would be able to make progress at very different rates even if both given identical compute resources.

At a boots-on-the-ground level, the speed of progress an AI project makes will be influenced by thosands of tiny decisions abou... (read more)

5
rosehadshar
I agree that it's not necessarily true that centralising would speed up US development! (I don't think we overlook this: we say "The US might slow down for other reasons. It’s not clear how the speedup from compute amalgamation nets out with other factors which might slow the US down: * Bureaucracy. A centralised project would probably be more bureaucratic. * Reduced innovation. Reducing the number of projects could reduce innovation.") Interesting take that it's more likely to slow things down than speed things up. I tentatively agree, but I haven't thought deeply about just how much more compute a central project would have access to, and could imagine changing my mind if it were lots more.

I don't really know... I'm suspect some kind of first-order utility calculus which tallies up the number of agents which are helped per dollar weighted according to what species they are makes animal welfare look better by large degree. But in terms of getting the world closer on the path of the "good trajectory", for some reason the idea of eliminating serious preventable diseases in humans feels like a more obvious next step along that path?

Not really a question but...  if you guys ever released a piece of merch that was insanely expensive but most of the cost went to charity (e.g. some special edition $3000 Mr Beast branded Tshirt where you guys give 3k to GiveDirectly for every unit sold), I'd wanna buy them for all my friends.

A priori, what is the motivation for elevating the very specific "biological requirment" hypothesis to the level of particular consideration? Why is it more plausible than than similarly prosaic claims like "consciousness requires systems operating between 30 and 50 degrees celsius" or "consciousness requires information to propegate through a system over timescales between 1 millisecond and 1000 milliseconds" or "consiousness requires a substrate located less than 10,000km away from the center of the earth"?

4
Bradford Saad
(I like the question and examples!) I take motivations for the biological requirement and for considering it to be empirical rather than a priori.  One motivation for the biological requirement is that, in the cases we know about, fine-grained differences in consciousness seem to be systematically and directly underpinned by biological differences. This makes the biological requirement more plausible than many other claims at the same level of specificity. While there isn’t a corresponding motivation for the temperature and timescale claims, there are related motivations: at least in humans, operating in those ranges is presumably required for the states that are known to systematically and directly vary with fine-grained differences in consciousness; going towards either end of the 30-50 C temperature range also seems to render us unconscious, which suggests that going outside the range would do so as well. Looking beyond the human case, I take it that certain animals operating outside the 30-50 C range makes the temperature claim less plausible than the biological requirement. Admittedly, if we widen the temperature range enough, the resulting temperature claim will be as plausible as the biological requirement. But the resulting temperature claim’s plausibility will presumably be inherited from claims (such as the biological requirement) that are more informative (hence more worthy of consideration) with respect to which systems are conscious. As for the distance claim, perhaps it would be plausible if one had Aristotelian cosmological beliefs! But I take it we now have good reason to think that the physical conditions that can exist on Earth can also exist far beyond it and that fundamental laws don’t single out Earth or other particulars for special treatment. Even before considering correlational evidence regarding consciousness, this suggests that we should find it implausible that consciousness depends on having a substrate within a certain distance fro

It seems a little weird to me that most of the replies to this post are jumping to the practicalities/logistics of how we should/shouldn't implement official, explicit, community-wide bans on these risky behaviours.

I totally agree with OP that all the things listsed above generally cause more harm than good. Most people in other cultures/communities would agree that they're the kind of thing which should be avoided, and most other people succeed in avoiding them without creatiing any explicit institution responsible for drawing a specific line between corr... (read more)

Sorry I might not have made my point clearly enough. By remaining anonymous, the OP has shielded themselves from any public judgement or reputational damage. Seems hypocritical to me given the post they wrote is deliberately designed to bring about public judgement and affect the reputation of Nick Bostrom.

So I'm saying "if OP thinks it's okay to make a post which names Nick and invites us all to make judgements about him, they should also have the guts to name themselves"

I really don't think the crux is people who disagree with you being unwilling to acknowledge their unconscious motivations. I fully admit that sometimes I experience desires to do unsavory things such as

- Say  something cruel to a person that annoys me
- Smack a child when they misbehave
- Cheat on my taxes
- Gossip about people in a negative way behind their backs
- Eat the last slice of pizza without offering it to anyone else
- Not stick to my GWWC pledge
- Leave my litter on the ground instead of carrying it to a bin
- Lie to a family member and say "I'm... (read more)

If EA community organisers are ending up isolated from everyone not involved in EA, that a really big problem!

4
Severin
Yep, I'm with Xavier here. The rule incentivizes community builders a bit to not make EA their only social bubble (which is inherently good I think). And it is not without workarounds, all of which cushion the addressed problem. For example, it encourages local community builders to hand over event facilitation to others more often. And if the rule is publicly known, participants can take a break from events that one leader leads to get around the rule. If participants don't know the rule, they'd get informed about its existence when they hit on an organizer. In either case, the consequence of even intentionally working around the rule would be taking it slow.

Also the claim that

"We'll end up as lonely, dispirited incels rowing our little boats around in circles, afraid to reach out, afraid to fall in love." 

Srikes me as patently false given myself and many people I know personally who engage with EA have partners from outside the EA community

The main reason I disagree is that to me it seems plainly obvious that it's far better for a community organiser's motivations to be related to earning respect/advancing their career/helping others, rather than their reason for participating in EA being so they can have more sex. This is because, if they're motivated by wanting to have more sex, then this predictably leads to more drama and more sexual harrassment.

I also don't think you did enough to back up the inference "lots of people are motivated by sex, therefore we should try to harness this, instea... (read more)

5
Xavier_ORourke
Also the claim that "We'll end up as lonely, dispirited incels rowing our little boats around in circles, afraid to reach out, afraid to fall in love."  Srikes me as patently false given myself and many people I know personally who engage with EA have partners from outside the EA community

Why would it be a problem for long-term community builders?

Anecdote: I used to help run a local university group in Australia. While helping run that group, I didn't try to date or sleep with attendees. Also while running that group, I met wonderful woman in a separate context who wasn't involved in the EA community, we entered into a relationship, and are now happily married and expecting a child.

I've also got lots of EA friends who've done community building in the past and are in really happy romantic relationships with spouses they met in a non-EA context as well.

3
pseudonym
Not sure, but my guess is the worry here is if many EA community builders end up having the vast majority of their social connections tied to the EA space this might be an issue?

If Bostrom is not entitled to protection from random people on the forum making judgments about him, why should it be any different for OP?

1
konrad
I didn't say Duncan can't judge OP. I'm questioning the judgment.

From where I sit -It's really hard to guess at all the details and relevant context of what's going on (which is why I feel a bit stupid commenting on it... but I guess I can't resist lol).

Is FHI the only org being subject to a hiring freeze? Or is the university/philosophy department cutting costs in many places? Are conflicts with the philosophy department basically FHI's fault? Or is the bureaucracy dysfunctional/unfriendly to FHI in ways which made it impossible to keep them happy without making other costly tradeoffs? If Nick steps down as director, i... (read more)

Are conflicts with the philosophy department basically FHI's fault?

I think this is a relevant question, but I don't think it's the whole question (not that you were claiming it was). As an outsider who has heard some stories and has some guesses, I would conjecture that the University is (at least sometimes) unreasonable and bureaucratic, but nevertheless, if you want to be a director of a university-affiliated research group, "managing the relationship with the university, even when they are being unreasonable" is absolutely a core competency of the job, and it's not one that Nick has had much luck with.

Relevant context is that FHI had a hiring freeze and was dying before the apology. And yes, it's only FHI. There was some kind of cutting corners to avoid bureaucracy and the department got mad. It's possible that with another leader FHI would be able to be rehabilitated in the department. Nobody can say for sure, but their best bet would be to go with someone willing to play by the book, and it would be odd for the department to have a grudge against new leadership.

My understanding is that this is indeed unique to FHI, unfortunately. This is maybe why FHI and GPI make for a compelling comparison - both are EA-affiliated, both are in the University of Oxford. While working with a University is never easy, GPI seems totally fine and indeed does continue to hire, run events, etc. FHI does not.

I don't know about the alternatives to Bostrom or how likely they would be to change the situation. Nathan makes a good point that perhaps prediction markets could play a role here. I generally think that, given I run an EA researc... (read more)

If Bostrom did step down as FHI director, who is likely to replace him? How confident are you that  a new director will succeed in resolving conflicts with the broader philosophy department?

I have very little direct experience with FHI (just a very brief internship) but from the outside it looks like FHI has produced some really amazing research while Bostrom has directed it.

Perhaps a good way to appraise whether FHI has been performing above/below par during Bostrom's directorship is to compare its output to a similar organisation such as Global Prio... (read more)

Mervin makes a great point that it is hard to compare GPI to FHI in general. But I also think comparing past FHI and past GPI is not the right way of thinking about it - instead we want to compare current/expected future FHI to current/expected future GPI. And the fact of the matter is quite clear that current/expected future GPI still can actually hire people, engage in productive research work, and maintain a relationship with the university whereas current/expected future FHI I think can best be described as "basically dead".

FHI and GPI have different aims. I think the better comparison would be FHI today vs. what FHI could be. Or FHI today vs. FHI before things went to pot.

All the people I know who have recently worked at/work at FHI want him to step down, they just are uncomfortable saying so directly. I agree that this is not a good place to raise the issue, but I'm not sure where is, given that. I hope this thread emboldens others to have a frank conversation, and it's possibly helpful to have information out there. I would really like FHI to survive, and it sounds like it could if he stepped down.

We should be careful to avoid dismissing a simple easy solution to a real problem because it might fail to solve an imaginary one. Do you really think the community currently has a problem with bosses pressuring their direct reports to help them move house? 

How do you know we don't live in a world where >90% of the problem is specifically due to people having sex / trying to have sex with each other? What would convince you that sex is the culprit, rather than interpersonal relationships in general?

In my oppinion - a very attractive compromise which many other cultures adopt is to keep everythign you love about the deep relationships except for the sex. People having sex with each other is uniquely prone to causing harm+drama+conflict.

I don't think we'll ever see a TIME article exposing the problem that someone in EA had too many people offer to help them move house, or that community events were filled with too much warmth and laughter, or that people offered too much emotional support to someone when they lost a parent.

More friendship and loyalty and support and love and fun and shared moments of vulnerability is fine! Just leave out the sex part!

I don't think this fixes all of it. For example, imagine someone describing being expected to load their boss' personal belongings into a moving truck, on a weekend, with pizza and beer for compensation.

Also, many people will want to participate in EA professionally but not socially, and the stronger the community is socially the harder that will be.

Which isn't to say that we should avoid doing nice things for each other and having fun together, but it doesn't free us from thinking about how people might feel pressured.

1[comment deleted]

Thanks! Now that SBF has been disavowed do you think EA still has a big problem with under-emphasising conflicts of interest?

I still think the best critiques benefit from being extremely concrete and that article could have had more impact if it spent less time on the high-level concept of "conflicts of interest" and more time explicitely saying "crypto is bad and it's a problem that so many in the community don't see this"

3
Radical Empath Ismam
I felt the article was pretty concrete in saying exactly that,"crypto is bad ...". It didn't strike me as high level/ abstract at all.

Dear authors - could you please provide at least one concrete example of a high-quality "deep critique" of Effective Altruism which you think was given inadequate consideration?

7
Radical Empath Ismam
I'm not the author, but there was a very prescient critique submitted to the EA criticism contest, that went underappreciated. https://medium.com/@sven_rone/the-effective-altruism-movement-is-not-above-conflicts-of-interest-25f7125220a5 UPDATE: actually I realised did specifically mention this critique as an example.

If there was no difference at all between the beliefs/values/behaviours of a the average member of this community, versus the average member of the human species - then there would be no reson for the concept "Effective Altruism" to exist at all.

It would be a terrible thing for our community to directly discriminate against traits which are totally irrelevant to what someone has offer to the EA project (such as race/gender/sexual preference) - and I've never heard anyone around here disagree with that.

But when it comes to a traits such as being highly inte... (read more)

This is consistent with the point I'm trying to make - all human interactions in all contexts are happening within a super complex web of norms and taboos, and any proposal as simple as "just let people do whatever they want" is a non-starter

I agree that if all the concerns about relationship norms in EA culture were comming from small minority, then this would not justify changing them (but the minority are still entitled to try and advocate/persuade). 

But when it comes to culture/status, I think the dynamics mean majority rule is pretty much baked in by default anyway! So we might not have to worry much about that.

Interestingly, we might have different impressions about what the median attitude is in the community when it comes to questions like 

- "Is it a bad idea to have sex with... (read more)

I think it's worth keeping in mind that every functioning culture includes a vast list of taboos which people enforce on others. In both regular Western culture, and EA culture, all of the following are considered inappropriate:

- Making sexist jokes
- Not being sympathetic when someone else experiences a personal tragedy
- Nuidity in the workplace
- Answering somebody else's phone
- Smoking weed in the office
- Bragging about money
- Chewing with your mouth open
- Refusing to shake somebody's hand when you're introduced
- Making prolonged, intense eye contact with... (read more)

7
Evan_Gaensbauer
In terms of what's considered appropriate in "regular western culture," a lot of this is not true enough to justify the generalizations you're making: * There are variations within cultures in any country, never mind between western, and all other countries, whereby the extent to which crude sexism is considered appropriate. I've met many men from some different walks of life just in Canada whose sense of what's normal is such that they'll look down on other men who don't tow the line with their chauvinistic attitudes and misogynistic comments. * While it's far from being all of them, there are a lot of sections of the upper class where bragging about how much money one makes is considered respectable, and this influences other aspects of culture too, especially in North America. * Making intense eye contact during a normal interaction is considered inappropriate in most cultures, though there is relative nuance here. Spending a longer amount of time making direct eye contact as part of a back-and-forth in conversation is much more accepted in western cultures, for example, compared to in Russia or China, to the point that to avoid too much making eye contact during conversation in western cultures is often considered rude.
4
Amber Dawn
So first, I do in fact want EA culture to lean substantially more towards personal autonomy than Western culture does - I like autonomy a lot! I think some regulation is inevitable as people interact. For example, if I go around calling everyone names and never showering, people might be like "Amber is smelly and rude, so I'm not going to invite her to my parties". And then, if I noticed people never invited me to parties, I might be like 'huh! maybe I should shower, and be less rude'. So in a way, people there are 'controlling' and 'regulating' my behaviour, but that doesn't seem overly coercive. What I object to is if one person thinks I'm smelly and rude and is like "no-one else should invite Amber to parties". This seems to me to be illegitimately hijacking the norm-creation process. Like ideally norms should arise out of the majority preferences of the group; it's bad if a minority decide This Is How It's Going To Be and the rest have to conform.   

Great point! As well as "focus on the norms that are causing the most harm" I'd want to also add  "focus on the norms that promise the least benefit".

Doing weird things like giving away 10% of your income, or talking about shrimp welfare, or raising the alarm about the dangers of AGI are all very weird, but there are credible worldviews in which it's really important to do them anyway. 

Whereas weird things like having sex with multiple people within your professional network on a regular basis promise mild benefits at best, even according to worldviews which endorse them.

Another important difference with monogomy is that it's taboo to make a proposition to somebody who's already married or already in a serious relationship, so people don't make them as often.

Pollyamory is not necessarily a bad thing in all contexts and all implementations, and I'm not claiming that everyone who practices is an abuser - but on its face it seems intuitive that the prevalance of polly in a community would interact with frequency of sexual harrasment/assult (especially when layered on top of other things like high prevalance of aspergers+mood dissorders+professional relationships between members of the community).

I'm not advocating this entirely, but just to illustrate the point - imagine if most people in EA had cultural attitude... (read more)

For what it's worth, I'm on the 1 year "try giving" pledge atm. Even though I currently donate to "EA approved" orgs, I would never make the life-long commitment if I thought it meant giving up the right to use my own discretion when deciding which causes/organisations to give to.

4
NickLaing
Love it nice one Xavier! I completely agree with you, and I don't see the pledge as even having to have a direct relationship to EA. I think anyone could legitimately sign the pledge outside of any direct EA framework, giving money to whatever was in your opinion was an effective organsation.

That's true, but before the brand "Effective Altruism" existed, there was no reason why starting an organisation using that name should have made the founders beholden to the will of every single participant in this community - you'd need to conjecture a pretty unreasonable amount of foresight and scheming to think that even back then the founders were trying to structure these orgs in a manner designed to maintain central control over the movement.

If you or me or anyone else wanted to start our own organisation under a new brand with similar goals to CEA or GWWC I don't think anyone would try to stop us!

7
Milan Griffes
My model is that no one would try to formally stop this effort (i.e. via a lawsuit), though it would receive substantial pushback in the form of:  * Private communication discouraging the effort  * Organizers of the effort excluded and/or removed from coordinating fora, such as EA slack groups  * Public writing suggesting that the effort be rolled into the existing EA movement  * Attempts (by professional EAs) to minimize the funding directed to the effort from traditional EA funders (i.e. the effort would be viewed as a competitor for funding) 
7
Milan Griffes
I don't follow what you're pointing to with "beholden to the will of every single participant in this community." My point is that CEA was established as a centralizing  organization to coordinate the actions and branding of the then-nascent EA community.  Whereas Luke's phrasing suggests that CEA drove the creation of the EA community, i.e. CEA was created and then the community sprung up around it. 

I encourage readers to consider whether they are the correct audience for this advice. As I understand it, this advice is directed at those for whom all of the following apply:

- Making a large impact on the world is overwhelmingly more important to you than other things people often want in their lives (such as spending lot of time with friends/family, raising children, etc.)
- You have already experienced a normal workload of ~38h per week for at least a couple of years, and found this pretty easy/comfortable to maintain
- You generally consider yourself to... (read more)

3
ramekin
If you're not the right person for the article, I'd instead recommend this post on Sustained effort, potential, and obligation. I've found it's given me a helpful framework for making sense of my own limits on working hours, and you may also find it useful. I admit that reading this post also stirred up some feelings of inadequacy for me—because, unlike all those CEOs and great men of history, I have actually a pretty low-to-average limit on how much sustained effort my brain will tolerate in a day. If you find yourself with similar feelings (which might be distressing, perhaps even leading you to spiral into self-hatred and/or seek out extreme measures to 'fix' yourself), the best antidote I've found for myself is The Parable of the Talents by Scott Alexander. (TLDR: variation in innate abilities is widespread, and recognizing and accepting the limits on one's abilities is both more truthful and more compassionate than denying them.)
2
Henry Stanley 🔸
Some stimulants seem to work well for depression, however: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6375494/

And it would probably be a huge mistake to seek out an adderall prescription.

...unless you have other reasons to believe that an Adderall prescription might be good for you. Saliently: if you have adhd symptoms.

Even if all those things apply ... this post may not be for you! Last year I tried to replace sleep with caffeine, and it did not go well. Even if you think you're happy and emotionally stable, you may discover that stimulants are anxiogenic for you, and you may be dumb enough (i.e., me) not to make that connection for a year. 

Stimulants, at least for me, are much better at making me feel productive than increasing my total output. I regularly wasted 3-6 hours chunks plumbing the depths of a rabbit hole that unstimulated me would have rightly avoided. A moderate caveat emptor here. 

Sorry Sophia but I still don't completely understand how what we're talking about maps on to actual decisions community builders are making. I'm still suspicious that many of us are sensing a vibe problem but misdiagnosing it as a messaging/cause prioritization problem.

I would find it really helpful if you could an example of how you percieve the big tent / small tent abstraction could map on to a concrete action which a community organiser takes. 

Let's say someone was starting up a new a  EA group at an Australian university - what's an example of a mistake you worry this person might make if they're too directly focused on chasing the "tails"?

1
Sophia
Relevant context: Xavier used to do lots of community building and helped run an Australian university EA group Yeah, I think vibes are a big deal, where vibes is pointing to something like "people have fun at events because the social dynamics feel good" (where "fun from social dynamics" is distinct from" fun from the intellectual stimulation of the philosophical puzzles" or other sources of fun).  Maybe this doesn't add anything to that/everything else that I would say I'm pointing to with the whole "median" targeting thing maybe also impacts vibes?  Campground/tent model is maybe more useful for understanding the role of vibes The campground/tent distinction helps me form my thoughts more explicitly around the role that vibes play (I do think they are not the only thing in a vague cluster of related strategies that I want to point to, but they are definitely a huge chunk/a lot of the other strategies maybe are important because of their impact on "vibes").  In short, our "tent" is the current movement and the "campground" is everyone who isn't in the movement (got a top-level post in the works, privately shared a draft with Xavier already but for anyone else who wants to give feedback before I post it, please DM me!).  Vibes at events seem important for the campground, ie. for all the people who aren't involved.  For example, a positive campground effect of vibes could go something like: Someone comes to an event who after has a great impression of effective altruism but they don't end up engaging much more, maybe because they just happen to get busy or it's not really quite their thing intellectually. Then, years later, they might go on to say nice things years later to their colleagues or friends who mention effective altruism. Those colleagues and friends can then engage with these ideas they agree with without being socially punished. The point of these models [1] The point of the above comments and also the campground/tent model is more about est

Thanks Holden and Luke for answering so many questions <3

Thinking in terms of broad generalisations/approximations - if you had to draw a graph depicting the value provided to OP by a new research analyst over time, what kind of shape would this graph have?

Or to ask the question in a different way: Are your efforts to hire for new OP roles motivated more by a desire to make better grants in the next couple of years, or by an intention to have a strong team in place several years from now which does high quality work in the future?

3
lukeprog
Responding to your second formulation of the question, the answer is "more the latter than the former." We intend to invest heavily in training and mentoring new hires, and we hope that research analysts will end up being long-term core contributors at Open Phil — as research analysts, as grant investigators, and as high-level managers, among other roles — or, in some cases, in important roles that require similar skills outside Open Phil.