I am an (aspiring) x-risk researcher, and president at EA Groningen for the past 2 years. I am especially interested in crucial considerations within longtermism.
I have a background in (moral) philosophy, business admin, and moral psychology.
Thanks for the elaborate reply!
I think there's a lot of open space in between sending out surveys and giving people binding voting power. I'm not a fan of asking people to vote on things they don't know about. However, I have something in mind of "inviting people to contribute in a public conversation and decision-making process". Final decision power would still be with CEA, but input is more than one-off, the decision-making is more transparant, and a wider range of stakeholders is involved. Obviously, this does not work for all types of decisions - some are too sensitive to discuss publicly. Then again, it may be tempting to classify many decisions as "too sensitive". Well, organisation "opening up" should be an incremental process, and I would definitely recommend to experiment with more democratic procedures.
Hi Max, good to read an update on CEA's plans.
Given CEA's central and influential role in the EA community, I would be interested to hear more on the approach on democratic/communal governance of CEA and the EA community. As I understand it, CEA consults plenty with a variety of stakeholders, but mostly anonymously and behind closed doors (correct me if I'm wrong). I see lack of democracy and lack of community support for CEA as substantial risks to the EA community's effectiveness and existence.
Are there plans to make CEA more democratic, including in its strategy-setting?
There will be a lot to learn from the current pandemic from global society. Which lesson would be most useful to "push" from EA's side?
I assume this question is in between the "best lesson to learn" and "lesson most likely to be learned". We probably want to push a lesson that's useful to learn, and that our push actually helps to bring it into policy.
Given the high uncertainty of this question, would you (Toby) consider giving imprecise credences?
Not a funding opportunity, but I think a grassroots effort to employ social norms to enforce social distancing could be effective in countries in early stages where authorities are not enforcing, e.g. The Netherlands, UK, US, etc.
Activists (Student EA's?) could stand with signs in public places asking people non-aggressively to please go home.
I think this article very nicely undercuts the following common sense research ethics:
If your research advances the field more towards a positive outcome than it moves the field towards a negative outcome, then your research is net-positive
Whether research is net-positive depends on the current field's position relative to both outcomes (assuming that when either outcome is achieved, the other can no longer be achieved). It replaces this with another heuristic:
To make a net-positive impact with research, move the field closer to the positive outcome than the negative outcome with a ratio of at least the same ratio as distance-to-positive : distance-to-negative.
If we add uncertainty to the mix, we could calculate how risk averse we should be (where risk aversion should be larger when the research step is larger, as the small projects probably carry much less risk to accidentally make a big step towards FAI).
The ratio and risk-aversion could lead to some semi-concrete technology policy. For example, if the distance to FAI and UAI is (100, 10), technology policy could prevent funding any projects that either have a distance-ratio (for lack of a better term) lower than 10 or that have a 1% or higher probability a taking a 10d step towards UAI.
Of course, the real issue is whether such a policy can be plausibly and cost-effectively enforced or not, especially given that there is competition with other regulatory areas (China/US/EU).
Without policy, the concepts can still be used for self-assessment. And when a researcher/inventor/sponsor assesses the risk-benefit profile of a technology themselves, they should discount for their own bias as well, because they are likely to have an overly optimistic view of their own project.
I really love Charity Entrepreneurship :) A remark and a question:
1. I notice one strength you mention at family planning is "Strong funding outside of EA" - I think this is a very interesting and important factor that's somewhat neglected in EA analyses because it goes beyond cost-effectiveness. We are not asking the 'given our resources, how can we spend them most effectively?' but the more general (and more relevant) 'how can we do the most good?' I'd like to see 'how much funding is available outside of EA for this intervention/cause area' as a standard question in EA's cost-effectiveness analyses :)
2. Is there anything you can share about expanding to two of the other cause areas: long-termism and meta-EA?
A consulting organisation aimed at EA(-aligned) organisations, as far as I'm aware: https://www.goodgrowth.io/.
Mark McCoy, mentioned in this post, is the Director of Strategy for it.
This might be just restrating what you wrote, but regarding learning unusual and valuabe skills outside of standard EA career paths:
I believe there is a large difference in the context of learning a skill. Two 90th-percentile quality historians with the same training would come away with very different usefulness for EA topics if one learned the skills keeping EA topics in mind, while the other only started thinking about EA topics after their training. There is something about immediately relating and applying skills and knowledge to real topics that creates more tailored skills and produces useful insights during the whole process, which cannot be recreated by combining EA ideas with the content knowledge/skills at the end of the learning process. I think this relates to something Owen Cotton-Barratt said somewhere, but I can't find where. As far as I recall, his point was that 'doing work that actually makes an impact' is a skill that needs to be trained, and you can't just first get general skills and then decide to make an impact.
Personally, even though I did a master's degree in Strategic Innovation Management with longtermism ideas in mind, I didn't have enough context and engagement with ideas on emerging technology to apply the things I learned to EA topics. In addition, I didn't have the freedom to apply the skills. Besides the thesis, all grades were based on either group assignments or exams. So some degree of freedom is also an important aspect to look for in non-standard careers.