MY

Misha_Yagudin

2219 karmaJoined Feb 2019

Comments
269

(From an email.) Some questions I am interested in:

  1. What's the size of alexithymia (A)?

  2. Does it actually make MH issues more likely or more severe? This mashes a few plausible claims and needs to be disentangled carefully, e.g., (a) given A, does MH more likely to be developed in the first place; (b) given MH, will A (even if acquired as a result of MH) make MH issues last longer or be worse? A neat casual model might be helpful here, separating A acquired with MH vs. A pre-existing to MH.

  3. How treatable is A? Does treating A improves MH? Is there any research (esp. RCT) on this? Does this depends on subgroups, e.g., A acquired with MH vs. A pre-existing to MH vs. A without MH…

  4. How would treating A fit into the MH treatment landscape? Can it be integrated with ongoing MH efforts in general (like people generally seeing therapy or doing CBT with a book or an app)? Can it be integrated with existing seemingly effective solutions (e.g., https://www.charityentrepreneurship.com/our-charities incubated MH orgs or charities recommended by https://www.happierlivesinstitute.org/)?

Yes, the mechanism is likely not alexithymia directly causing undesirable states like trauma but rather diminishing one's ability to get unstack given that traumatic events happened.

Ah, I didn't notice the forecasting links section... I was thinking of adding a hyperlink to the question to the name of the platform at the highlighted place.

Also, maybe expanding into the full question when you hover over the chart? 

 

I think this is great!

https://funds.effectivealtruism.org/funds/far-future might be a viable option to get funding.

As for suggestions,

  • maybe link to the markets/forecasting pools you use for the charts like this "… ([Platform] (link-to-the-question))?
  • I haven't tested, but it would be great for links to your charts to have snappy social media previews.

If you think there is a 50% chance that your credences will say go from 10% to 30%+. Then you believe that with a 50% probability, you live in a "30%+ world." But then you live in at least a 50% * 30%+ = 15%+ world rather than a 10% world, as you originally thought.

FWIW, different communities treat it differently. It's a no-go to ask for upvotes at https://hckrnews.com/ but is highly encouraged at https://producthunt.com/.

Good luck; would be great to see more focus on AI per item 4!

Is Rebecca still a fund manager, or is the LTFF page out of sync?

So it's fair to say that FFI-supers were selected and evaluated on the same data? This seems concerning. Specifically, on which questions the top-60 were selected, and on which questions the below scores were calculated? Did these sets of questions overlap?

The standardised Brier scores of FFI superforecasters (–0.36) were almost perfectly similar to that of the initial forecasts of superforecasters in GJP (–0.37). [17] Moreover, even though regular forecasters in the FFI tournament were worse at prediction than GJP forecasters overall (probably due to not updating, training or grouping), the relative accuracy of FFI's superforecasters compared to regular forecasters (-0.06), and to defence researchers with access to classified information (–0.1) was strikingly similar.[18]

Load more