All of NthOrderVices's Comments + Replies

Here's a relevant set of estimates from a couple of years ago, which has a guesstimate model you might enjoy. Your numbers seem to be roughly consistent with theirs. They were trying to make a broader argument that "1. EA safety is small, even relative to a single academic subfield. 2. There is overlap between capabilities and short-term safety work. 3. There is overlap between short-term safety work and long-term safety work. 4. So AI safety is less neglected than the opening quotes imply. 5. Also, on present trends, there’s a good chance that academia wi... (read more)

I'm not sure what campus EA practices are like - but, in between pamphlets and books, there are zines. Low-budget, high-nonconformity, high-persuasion. Easy for students to write their own, or make personal variations, instead of treating like official doctrine. ie, https://azinelibrary.org/zines/

Nice. And when it comes to links, ~half the time I'll send someone a link to the Wikipedia page on EA or longtermism rather than something written internally.

What are some practical/theoretical developments that would make your work much less/more successful than you currently expect? (4 questions in 1, but feel free to just answer the most salient for you)

2
Stuart Armstrong
2y
Most of the alignment research pursued by other EA groups (eg Anthropic, Redwood, ARC, MIRI, the FHI,...) would be useful to us if successful (and vice versa: our research would be useful for them). Progress in inner alignment, logical uncertainty, and interpretability is always good. Fast increase in AI capabilities might result in a superintelligence before our work is ready. If the top algorithms become less interpretable than they are today, this might make our work harder. Whole brain emulations would change things in ways that are hard to predict, and could make our approach either less or more successful.

To some limited degree, some people have some beliefs that are responsive to the strength of philosophical or scientific arguments, and have some actions that are responsive to their beliefs. That's about as weak a claim as you can have without denying any intellectual coherence to things. So then the question becomes, is that limited channel of influence enough to drive major societal shifts?

 Or actually, there might be two questions here: could an insight in moral philosophy alone drive a major societal shift, so that society drifts towards whichever argument is better? and, to what extent has actual moral progress been caused by intellectual catalysts like that?