This is also great, thank you!
Not the intended audience, but as a US person who lives in the Bay Area, I enjoyed reading this really detailed list of what's often unusual or confusing to people from a specific different cultural context
I generally directionally agree with Eli Nathan and Habryka's responses. I also weak-downvoted this post (though felt borderline about that), for two reasons.
(1) I would have preferred a post that tried harder to even-handedly discuss and weigh up upsides and downsides, whereas this mostly highlighted upsides of expansion, and (2) I think it's generally easier to publicly call for increased inclusivity than to publicly defend greater selectivity (the former will generally structurally have more advocates and defenders). In that context I feel worse about (1) and wish Scott had handled that asymmetry better.
But I wouldn't have downvoted if this had been written by someone new to the community, I hold Scott to a higher standard and I'm pretty uncertain about the right policy with respect to voting differently in response to the same content on that basis.
Quite. I was in that Stanford EA group, I thought Kelsey was obviously very promising and I think the rest of us did too, including when she was taking a leave of absence.
No worries, appreciate ppl checking :)
As noted in the post, I got Scott's permission before posting this.
I strongly disagree with Greg. I think CFAR messed up very badly, but I think the way they messed up is totally consistent with also being able to add value in some situations.
We have data I find convincing suggesting a substantial fraction of top EAs got value from CFAR. ~ 5 years have passed since I went to a CFAR workshop, and I still value what I learned and think it's been useful for my work. I would encourage other people who are curious to go (again, with the caveat that I don't know much about the new program), if they feel like they're in a place of relative strength and can take a discerning eye to what they're taught.
If I, with (mostly) admirable candour, describe a series of grossly incompetent mistakes during my work as a doctor, the appropriate response may still be to disqualify me from future medical practice (there are sidelines re. incentives, but they don't help)
I think doctor is a really disanalogous example to use; doctors are in one of the relatively few professions where screwups regularly lead to death; we want to some somewhat risk-averse, with respect to doctors (and e.g. pilots or school bus drivers), at least if the screwups are the very dangerous kind (as opposed to like, being terrible at filing one's paperwork), and aren't based on a reasonable CBA (e.g. enrolling patients in a clinical trial with a drug that looked promising but turned out to be dangerous). For lots of other professions, this example looks way less compelling; e.g. I doubt people would think that a startup founder or movie director or author who had a bunch of failures but also some big wins should be banned from their profession or ostracized in their community. I think in-person overnight events about psychology are in a pretty in-between risk category.
You said you wouldn’t tell anyone about your friend’s secret, but this seems like a situation where they wouldn’t mind, and it would be pretty awkward to say nothing…etc.
This isn't your main point, and I agree there's a lot of motivated cognition people can fall prey to. But I think this gets a bit tricky, because people often ask for vague commitments, that are different from what they actually want and intend. For example, I think sometimes when people say "don't share this" they actually mean something more like "don't share this with people that know me personally" or "keep it in our small circle of trusted friends and advisors" or "you can tell your spouse and therapist, but no one else" (and often, this is borne out when I try to clarify). Sometimes, I think they are just trying to convey "this info is sensitive, tread with care". Or, they might mean something more intense, like "don't share this, and aim not to reveal any information that updates others substantially towards thinking it's true".
Clarification can often be useful here (and I wish there were more verbal shorthands for different levels of intensity of commitment) but sometimes it doesn't happen and I don't think, in its absence, all agreements should be taken to be maximally strict (though I think it's extremely important to have tools for conveying when a requested agreement is very strict, and being the kind of person that can honor that). And I think some EAs get intense and overly scrupulous about obeying unimportant agreements, which can be pretty unwieldy and divorced from what anyone intended.
I think "do you keep actually-thoughtful promises you think people expected you to interpret as real commitments" and "do you take all superficially-promise-like-things as serious promises" are fairly different qualities (though somewhat correlated), and kinda often conflated in a way that I think is unhealthy and even performative.
This seems really exciting, and I agree that it's an underexplored area. I hope you share resources you develop and things you learn to make it easier for others to start groups like this.PSA for people reading this thread in the future: Open Phil is also very open to and excited about supporting AI safety student groups (as well as other groups that seem helpful for longtermist priority projects); see here for a link to the application form.
I used to agree more with the thrust of this post than I do, and now I think this is somewhat overstated.
[Below written super fast, and while a bit sleep deprived]
An overly crude summary of my current picture is "if you do community-building via spoken interactions, it's somewhere between "helpful" and "necessary" to have a substantially deeper understanding of the relevant direct work than the people you are trying to build community with, and also to be the kind of person they think is impressive, worth listening to, and admirable. Additionally, being interested in direct work is correlated with a bunch of positive qualities that help with community-building (like being intellectually-curious and having interesting and informed things to say on many topics). But not a ton of it is actually needed for many kinds of extremely valuable community building, in my experience (which seems to differ from e.g. Oliver's). And I think people who emphasize the value of keeping up with direct work sometimes conflate the value of e.g. knowing about new directions in AI safety research vs. broader value adds from becoming a more informed person and gaining various intellectual benefits from practice engaging with object-level rather than social problems.
Earlier on in my role at Open Phil, I found it very useful to spend a lot of time thinking through cause prioritization, getting a basic lay of the land on specific causes, thinking through what problems and potential interventions seemed most important and becoming emotionally bought-in on spending my time and effort on them. Additionally, I think the process of thinking through who you trust, and why, and doing early audits that can form the foundation for trust, is challenging but very helpful for doing EA CB work well. And I'm wholly in favor of that, and would guess that most people that don't do this kind of upfront investment are making an important mistake.
But on the current margin, the time I spend keeping up with e.g. new directions in AI safety research feels substantially less important than spending time on implementation on my core projects, and almost never directly decision-relevant (though there are some exceptions, e.g. I could imagine information that would (and, historically, has) update(d) me a lot about AI timelines, and this would flow through to making different decisions in concrete ways). And examining what's going on with that, it seems like most decisions I make as a community-building grantmaker are too crude to be affected much by additional info at the relevant level of granularity intra-cause, and when I think about lots of other community-building-related decisions, the same seems true.
For example, if I ask a bunch of AI safety researchers what kinds of people they would like to join their teams, they often say pretty similar versions of "very smart, hardworking people who grok our goals, who are extremely gifted in a field like math or CS". And I'm like "wow , that's very intuitive, and has been true for years, without changing". Subtle differences between alignment agendas do not, in my experience, bear out enough in people's ideas about what kinds of recruitment are good that I've found it to be a good use of time to dig in on. This is especially true given that places where informed, intelligent people who have various important-to-me markers of trustworthiness differ are places where I find that it's particularly difficult for an outsider to gain much justified confidence.
Another testbed is that I spend a few years spending a lot of time on Open Phil's biosecurity strategy, and I formed a lot of my own, pretty nuanced and intricate views about it. I've never dived as deep on AI. But I notice that I didn't find my own set of views about biosecurity that helpful for many broader community-building tradeoffs and questions, compared to the counterfactual of trusting the people who seemed best to me to trust in the space (which I think I could have guessed using a bunch of proxies that didn't involve forming my own models of biosecurity) and catching up with them or interviewing them every 6mo about what it seems helpful to know (which is more similar to what I do with AI). Idk, this feels more like 5-10% of my time, though maybe I absorb additional context via osmosis from social proximity to people doing direct work, and maybe this helpful in ways that aren't apparent to me.