N

nikos

555 karmaJoined May 2021

Comments
22

nikos
2mo32

Interesting, thanks for sharing the paper. Yeah agree that using the Brier score / log score might change results and it would definitely be good to check that as well. 

nikos
2mo10

In principle yes. In practice also usually yes, but the specifics depend on whether the average user who predicted on a question gets a positive amount of points. So if you predicted very late and your points are close to zero, but the mean number of points forecasters on that question received is positive, then you will end up with a negative update to your reputation score. 
Completely agree that a lot hinges on that reputation score. It seems to work decent for the Metaculus Prediction, but it would be good to see what results look like for a different metric of past performance. 

nikos
2mo30

Not sure how to quantify that (open for ideas). But intuitively I agree with you and would suspect it's at least a sizable part

nikos
2mo20

Yeah, definitely. The title was a bit tongue-in-cheek (it's a movie quote)

nikos
3mo20

And is the code to the MetaculusBot public somewhere? :) 

nikos
3mo32

It should be possible to fully automate the bot and just run a CRON job that regularly checks the Metaculus API for new questions, right? 

nikos
3mo10

I slightly tend towards yes, but that's mere intuition. As someone on Twitter put it, "Metaculus has a more hardcore user base, because it's less fun" - I find it plausible that the Metaculus user base and the Manifold user base differs. But higher trading volume I think would have helped. 

For this particular analysis I'm not sure correcting for the number of forecasters would really be possible in a sound way. It would be great to get the MetaculusBot more active again to collect more data. 

nikos
3mo50

Is it possible to get rid of the question mode for this post?

nikos
4mo10

For Metaculus there are lots of ways to drive engagement: prioritise making the platform easier to use, increase cash prizes, community building and outreach etc. 

But as mentioned in the article the problem in practice is that the bootstrap answer is probably misleading, as increasing the number of forecasters likely changes forecaster composition. 

However, one specific example where the analysis might be actually applicable is when you're thinking about how many Pro Forecasters you hire for a job. 

nikos
4mo10

In principle yes, you'll just still always have the problem that people are predicting at different time points. If the best and the 2nd best predict weeks or months apart then that changes results. 

Load more