Rockwell

Director @ EA NYC
4653 karmaJoined Aug 2021Working (6-15 years)
effectivealtruism.nyc

Bio

Participation
5

  • Full-time Director for EA NYC
  • Feel free to contact me by email for any questions or collaboration ideas or book a time in my Calendly.

Sequences
1

EA in NYC

Comments
107

Topic contributions
3

I'm really excited to see this! Thank you for all the work you've put into it!

One piece of feedback (that others in the actual target audience are probably better to weigh in on): I noticed the people page currently seems to exclusively feature white Americans and Europeans, which I imagine might be offputting to some people the website is designed to reach.

I'd like to second Ben and make explicit the concern about platforming ideologues whose public reputation is seen as pro-eugenics.

The best option for you will be very individual! I recommend taking the Animal Justice Academy course for an overview of types of advocacy in the animal movement and the Animal Advocacy Careers course if you are considering a career transition (now or in the future). If you are in/near the US, the AVA Summit in May is also a great way to dive in.

Without commenting on the rest of this case or EA Funds more broadly, this stood out to me:

At the EA funds website, they write that they usually grant money within 21 days from sending an application, and that their managers care (no further specification).

I was surprised the OP would request a response within one month when applying for a grant until I saw this truly is emphasized on the EA Funds site. This seems inconsistent with my understanding of many people's experiences with EA Funds and easy messaging to change to set more realistic expectations. I appreciate EA funders' efforts toward quick turnaround times, but traditional funders typically take many months to reach a decision, even for comparably sized (i.e. small) grants. This seems like a strong case for "underpromise, overdeliver."

I think that's a common intuition! I'm curious if there were particular areas covered (or omitted) from this post that you see as more clearly the natural function of one versus the other.

I'll note that a couple factors seem to blur the lines between city and national MEARO functions:

-Size of region (e.g. NYC's population is about 8 million, Norway's is about 5.5 million)
-Composition of MEAROs in the area (e.g. many national MEAROs end up with a home base city or grew out of a city MEARO, some city MEAROs are in countries without a national MEARO)

I could see this looking very different if more resources went toward assessing and intentionally developing the global MEARO landscape in years to come.

Thank you for writing and sharing this, Alix! I'm sorry that it was scary for you to post and I'm glad you did. You also linked to so many other useful readings I hadn't seen previously!

I'm wondering how these dynamics play out across different platforms and spaces—e.g. hiring processes for organizations with varying degrees of international staff vs. international online platforms like the Forum or EA Anywhere Slack vs. in-person events—and if there are better moderation mechanisms for acknowledging and accounting for language barriers across each. Online, for example, it's easy to list the languages you speak and some organizations list this on their staff pages (e.g. "You can contact Alix in French and English."). Maybe this could be added to Forum profiles or EA Global Swapcard profiles.

I'm also wondering how we can better account for this as community builders, especially in places with many immigrants. We remind attendees at the start of most EA NYC events that everyone present has a different starting point and we all have something to learn and something to teach. We began doing this, in large part, to make sure newcomers who don't "speak EA" feel welcome. But there might be a benefit to also explicitly noting possible language barriers, given how deeply international and multicultural the community here is. This is also making me want to look into facilitation trainings specifically focused on these dynamics; I'm sure there are non-obvious things we could be doing better.

@abrahamrowe, I'm curious if you have insights on the larger point about good governance across the EA ecosystem. As evidenced by EV's planned disbanding, sponsorship arrangements have a higher potential to become fraught. The opacity of the relationship between Rethink Charity and Nonlinear might be another example. (I.e. This is further indication Nonlinear employees wouldn't have had the same protection and recourse mechanisms as employees of more conventionally governed 501c3s, especially those of established 501c3s sizeable enough to hire 21 staff members.) Given RP is growing into one of the larger fiscal sponsors through your Special Projects Team, it might be worth further commentary from the RP team on how you're navigating risk and responsibility in sponsorship arrangements. Given RP's track record of proactive risk mitigation, I imagine you all have given this ample thought and it might serve as a template for others.

Once again, where is the board?

Two of the biggest questions for me are whether or not Nonlinear had a board of directors when Alice and Chloe worked for them and, if they did, whether an employee would know the identities and contact information of the board members and could feel reasonably safe approaching board members to express concerns and seek intervention. I can't find evidence they had a board at the time of the complaints or do now a year and a half after Alice and Chloe stopped working with them. The only reference to a board of directors I see in the Google Doc is Lightcone's board, which seems telling on a few levels.

Nonprofit boards are tasked with ensuring legal compliance, including compliance with relevant employment law considerations, and including above board practices in unconventional and riskier structures like Nonlinear chose to operate through. This situation looks very different if a legitimate board is in place than if employees don't have that safeguard.

Though I'm sad about the hurt experienced by many people across the Nonlinear situation, I'm personally less concerned with the minutiae of this particular organization and more about what structures, norms, and safeguards can be established across the EA ecosystem as a whole to reduce risk and protect EA community members going forward. Boards and institutional oversight are a recurring theme, from FTX to Nonlinear (to maybe OpenAI?) and I'm personally more skeptical of any organization that does not make its board information readily apparent.

Makes sense, thank you! Maybe my follow-up questions would be: How confident would they need to be that they'd use the experience to work on biorisk vs. global health before applying to the LTFF? And if they were, say, 75:25 between the two, would EAIF become the right choice -- or what ratio would bring this grant into EAIF territory?

Scattered first impressions:

  • I feel generally very positively about this update and have personally felt confused about the scope of EAIF when referring other people to it.
  • There are wide grey areas when attempting to delineate principles-first EA from cause-specific EA and the effective giving examples in this post stand out to me as one thorny area. I think it may make sense not to fund an AI-specific or an animal-specific effective giving project through EAIF (and the LTTF and AWF are more appropriate), but an effective giving project that e.g. takes a longtermist approach or is focused on near-term human and nonhuman welfare seems different to me. Put differently: How do you think about projects that don't cover all of EA, but also aren't limited to one cause area?
  • For this out-of-scope example in particular, I'm not sure where I would route someone to pursue alternative funding in a timely fashion:

Funding a very promising biology PhD student to attend a one-month program run by a prestigious US think tank to understand better how the intelligence community monitors various kinds of risk, such as biological threats ($6,000)

Maybe Lightspeed? But I worry there isn't currently other coverage for funding needs of this sort.

  • I'm worried about people couching cause-specific projects as principles-first, but there is already a heavy tide pushing people to couch principles-first projects as x-risk-specific, so this might not be a concern.
  • I'm really happy to see you thinking about digital minds and (seemingly) how to grow s-risk projects.
Load more