Rockwell

Co-Director @ EA NYC
4754 karmaJoined Working (6-15 years)
effectivealtruism.nyc

Bio

Participation
5

  • Full-time Director for EA NYC
  • Feel free to contact me by email for any questions or collaboration ideas or book a time in my Calendly.

Sequences
1

EA in NYC

Comments
115

Topic contributions
3

This seems like it has significant implications for the trajectory of global geopolitical stability and some GCR scenarios. I'm wondering whether or not others following this who are better informed than I am see this as a notable update.

Oh, it sounds like you might be confused about the context I'm talking about this occurring in, and I'm not sure that explaining it more fully is on-topic enough for this post. I'm going to leave this thread here for now to not detract from the main conversation. But I'll consider making a separate post about this and welcome feedback there.

I do also want to clarify that I have no desire to "control which ideas and people [anyone] is exposed to." It is more so, "If I am recommending 3 organizations I think someone should connect with, are there benefits or risks tied to those recommendations."

I really appreciate you sharing your perspective on this. I think these are extremely hard calls, as evidenced by the polarity of the discussion on this post, and to some extent it feels like a lose-lose situation. I don't think these decisions should be made in a vacuum and want other people's input, which is one reason I'm flagging how this affects my work and the larger involvement funnels in EA.

Thanks for spelling this out.

I think to give some color to how this affects my work in particular (speaking strictly for myself as I haven't discussed this with others on my team):

One of our organization priorities is ensuring we are creating a welcoming and hopefully safe community for people to do good better, regardless of people's identities. A large part of our work is familiarizing people with and connecting them to organizations and resources, including ones that aren't explicitly EA-branded. We are often one of their first touch points within EA and its niches, including forecasting. Many factors shape whether people decide to continue and deepen their involvement, including how positive they find these early touch points. When we're routing people toward organizations and individuals, we know that their perception of our recommendations in turn affects their perception of us and of EA as a whole.

Good, smart, ambitious people usually have several options for professional communities to spend their time within. EA and its subcommunities are just one option and an off-putting experience can mean losing people for good.

With this in mind, I will feel much more reluctant to direct community members to Manifold in particular and (EA-adjacent) forecasting spaces more broadly, especially if the community member is an underrepresented group in EA. I think Manifold brings a lot of value, but I can't in good conscience recommend they plug into communities I believe most people I am advising would find notably morally off-putting.

This is of course a subjective judgement call, I understand there are strong counterarguments here, and what repels one person also attracts another. But I hope this gives a greater sense of the considerations/trade-offs I (and probably many others) will have to spend time thinking about and reaching decisions around as a result of Manifest.

Austin, I'm gathering there might be a significant and cruxy divergence in how you conceptualize Manifold's position in and influence on the EA community and how others in the community conceptualize this. Some of the core disagreements discussed here are relevant regardless, but it might help clarify the conversation if you describe your perspective on this.

Commenting just to encourage you to make this its own post. I haven't seen a (recent) standalone post about this topic, it seems important, and though I imagine many people are following this comment section it also seems easy for this discussion to get lost and for people with relevant opinions to miss it/not engage because it's off-topic.

You might be interested in the information security sphere, which some in the EA community focus on, especially in the context of safe AI. This 80,000 Hours podcast from 2022 is a good overview: https://80000hours.org/podcast/episodes/nova-dassarma-information-security-and-ai-systems/

I'm really excited to see this! Thank you for all the work you've put into it!

One piece of feedback (that others in the actual target audience are probably better to weigh in on): I noticed the people page currently seems to exclusively feature white Americans and Europeans, which I imagine might be offputting to some people the website is designed to reach.

I'd like to second Ben and make explicit the concern about platforming ideologues whose public reputation is seen as pro-eugenics.

Load more