Pre-committing to not elaborating further, but I wanted to echo what is said in this comment and give a non-anonymous account as someone who (due to personal experience with reporting misconduct) also has similar feelings as KnitKnack.
Edit: I think Chana's comment is helpful context i.e. it seems good if people's expectations going in are calibrated to CEA CH's position is that it is there to "address problems that could prevent the effective altruism community from fulfilling its potential for impact".
In particular, they "don’t see pursuing justice as [their] mission" and "protecting people from bullies is sometimes a part of [their] work, and something [they'd] always like to be able to do, but it’s not [their] primary goal".
On a personal note, my advice to people who are considering going to CEA CH is to keep this in mind. To the extent to which there is a trade-off between impact and justice, it may not resolve in a way that is "just" from your POV, and their work on interpersonal harm does take the talent bottleneck seriously e.g. you should probably think about what they perceive the potential impact of the perpetrator to be.
+1. Did the January edition of this, and wholeheartedly endorse it! I think it was the best possible use of my time (and at least a few times better than the counterfactual), and it was incredibly well run (both with respect to the organisers and the mentors).
Thanks for curating the post!
Just a quick comment to highlight the responses which we have given to the list of disagreements, and to tweak your summary a bit to better reflect what I (not to speak for my other two co-authors) see our post as saying:
A good outside view is that markets are a good way of finding an outside view on a topic — in this case, on transformative AI. Long-term real rates would be higher than they currently are if markets believed that transformative AI was coming in the next ~30 years. If you believe timelines are short, you should personally be saving less or borrowing more. If you believe timelines are short and the market will realise a meaningful amount of time before transformative AI arrives, you should take a short Treasuries position. If you believe that the market should have already realised it and priced it in right now, you should rethink your timelines.
Edit: As it turns out, there's a nice third party summary which even more concisely captures the essence of what we are trying to get across!
To try to group/summarize the discussion in the comments and offer some replies:
1. ‘Traders are not thinking about AGI, the inferential distance is too large’; or ‘a short can only profit if other people take the short position too’
(a) Anyone who thinks they have an edge in markets thinks they've noticed something which requires such a large inferential distance that no one else has seen it.
(b) Many financial market participants ARE thinking about these issues.
The prospect of AGI is not a Thielian secret.
(c) Do make sure to read section X on “Trade risk and foom risk”, where we acknowledge that if you are both (i) extremely skeptical of market efficiency, and (ii) think foom is the likely takeoff scenario, then trading seems less like a good idea.
2. Stocks versus bonds
3. Other empirical evidence on real rates