- I frequently hear people say EAs rely too much on quantifying uncertain variables, but I basically never hear the opposite criticism. If everyone believes you shouldn't quantify, then nobody's doing it, so it can't possibly be true that people quantify too much, and in fact the opposite is probably true.
- Obviously I could make various counterarguments, like maybe the people who think we don't quantify enough are not writing essays about how we need to quantify more. Generally speaking, I don't think this counterargument is correct, but arguing for/against it is harder so I don't have much to say about it
- It's like Lake Wobegon, where all the children are above average. It's impossible for every single person in the community to believe that the community is not X enough
- Another example: everyone says we need to care more about systemic change
- Saw a Twitter post "EAs way under-update on thought experiments" and I thought, damn that's a spicy take. Then I realized I misread it and they actually said "over-update" and I thought...wow what a boring take that's been said a thousand times already
- They gave the simulation argument and Roko's Basilisk as examples. As far as I know, nobody has ever changed their behavior based on either of those arguments. It would be pretty much impossible for people to update less on them than they have
- I'm sure there are some people somewhere who have updated based on the simulation argument but I've never met them
- "People under-update on thought experiments" would have been a much more interesting take because people basically don't update on thought experiments
- They gave the simulation argument and Roko's Basilisk as examples. As far as I know, nobody has ever changed their behavior based on either of those arguments. It would be pretty much impossible for people to update less on them than they have
- By a shocking coincidence, I take the opposite side on all these examples: I think EAs should use more quantitative estimates, should care less about systemic change, and should update more on thought experiments
- Are there any issues where I make the same criticism as everyone else, and I'm actually wrong? Probably, idk
- I can think of some non-EA-related examples of this phenomenon, but I'm not as interested in those
- By analogy, the moment when the most people agree the stock market is going to go up is the exact moment when the market is at its peak. The price can't go higher because there's no one left to buy from. If everyone agrees, everyone /must/ be wrong
Relevant Scott Alexander: https://slatestarcodex.com/2014/03/24/should-you-reverse-any-advice-you-hear/ and https://slatestarcodex.com/2017/04/07/yes-we-have-noticed-the-skulls/. He said it better than me, but my post isn't about exactly the same thing so I figured it might be worth publishing.
(Note: The way I usually write essays is by writing outlines like this, and then fleshing them out into full posts. For a lot of the outlines I write, like this one, I never flesh them out because it doesn't seem worth the time. But I figured for Draft Amnesty Day, I could just publish my outline, and most people will get the idea.)
I think this outline needs major revisions to be improved.
The above is the clearest example of why I think this post's argument fails. It is definitely possible for all members to believe something about the community is inadequate. For example, say a sports team is bad at defence. And, every member of the team could believe that they need to improve their defence. The fact that all members believe it does not disprove the empirical fact that the team's defence is inadequate.
Where this impossibility claim might have legs is where the group belief is about group belief itself. For example, it may be impossible for every member of the team to believe that every member of the team does not think about defence enough. But the examples you talked about are not like this - they concern actual action, not group belief.
You are conflating noticing/talking/writing about an action with the action itself. This is especially apparent for issues that are larger than individual action. For the example of systemic change: every member could believe sincerely that the community as a whole should 'do more work on systemic change', but reasonably continue their normal, everyday, non-systemic work. In that case, everyone agrees that more systemic change work is needed, but no systemic change work actually gets done.
I think this is fairly common in EA counter-criticism, where people point to an old blog post about issue X, proving that EAs are already aware about X, and so, they argue, the criticism fails. While relevant, awareness of X pales in comparison to actually dealing with X itself.
This argument is further weakened by the fact that few critical stances are endorsed by near-100% of the community. There are significant counter-parties to most interesting critical claims. So making a critical claim is usually not in the situation of 'everyone already agrees with this'.
Finally, there's the common sense rebuttal that if a criticism is being made by many, that should (all else equal) increase your credence that the criticism is true. Contrarianism for contrarianism's sake is useful for checks and balances, but as a personal strategy is antithetical to epistemic modesty.
Yeah, it's mostly a heuristic argument, and the best you can do might be to just carefully look at the object level instead of trying to infer based on what people are saying.