All of Matt_Lerner's Comments + Replies

I'm also strongly interested in this research topic — note that although the problem is worst in the U.S., the availability and affordability of fentanyl (which appears to be driving OD deaths) suggests that this could easily spread to LMICs in the medium-term, suggesting that preventive measures such as vaccines could even be cost-effective by traditional metrics.

1
mincho
3mo
totally agree -- i think fentanyl is rightly understood as a huge a new threat, but i dont think there's a realization generally that fentanyl is essentially a technological advancement.  much stronger, much smaller, cheaper.  makes efforts to prevent drug trafficking much harder and makes harm reduction and social interventions much more difficult as well.  we beat cigarettes largely with price increases, fentanyl is a price decrease.  also it has shorter half life than heroin so people use it more often every day, which creates all sorts of other risks.  all of this is to say-- yes, fentanyl seems more likely to spread to countries that have been ok so far.

Easily reconciled — most of our money moved is via advising our members. These grants are in large part not public, and members also grant to many organizations that they choose irrespective of our recommendations. We provide the infrastructure to enable this.

The Funds are a relatively recent development, and indeed some of the grants listed on the current Fund pages were actually advised by the fund managers, not granted directly from money contributed to the Fund (this is noted on the website if it's the case for each grant). Ideally, we'd be able to gro... (read more)

We (Founders Pledge) do have a significant presence in SF, and are actively trying to grow  much faster in the U.S. in 2024.

A couple weakly held takes here, based on my experience:

  • Although it's true that issues around effective giving are much more salient in the Bay Area, it's also the case that effective giving is nearly as much of an uphill battle with SF philanthropists as with others. People do still have pet causes, and there are many particularities about the U.S. philanthropic ecosystem that sometimes push against individuals' willingness to t
... (read more)
5
Linch
6mo
It's great that you have a presence in SF and are trying to grow it substantially in 2024! That said, I'm a bit confused about what Founders' Pledge does; in particular how much I should be thinking about Founders' Pledge as a fairly GCR-motivated organization vs more of a "broad tent" org more akin to Giving What We Can or even the Giving Pledge. In particular, here are the totals when I look at your publicly-listed funds: * Climate Change ($9.1M) * Global Catastrophic Risks ($5.3M in 7 grants) * $3M of which went to NTI in October 2023. Congrats on the large recent grant btw! * Global Health and Development ($1.3M) * Patient Philanthropy Fund (~0) * Though to be fair that's roughly what I'd expect from a patient fund. From a GCR/longtermist/x-risk focused perspective, I'm rather confused about how to reconcile the following considerations for inputs vs outputs: * Founders' Pledge being around for ~7 years. * Founders' Pledge having ~50 employees on your website (though I don't know how many FTEs, maybe only 20-30?) * ~$10B(!) donations pledged, according to your website. * ~$1B moved to charitable sector * <20M total donations tracked publicly * <10 total grants made (which is maybe ~1.5-2 OOMs lower than say EA Funds) Presumably you do great work, otherwise you wouldn't be able to get funding and/or reasonable hires. But I'm confused about what your organizational mandate and/or planned path-to-impact is. Possibilities:  * You have a broad tent strategy aiming for greater philanthropic involvement of startup founders in general, not a narrow focus on locally high-impact donations * Founders' Pledge sees itself as primarily a research org with a philanthropic arm attached, not primarily a philanthropic fund that also does some research to guide giving * A very large fraction of your money moved to impactful charities is private/"behind the scenes", so your public funds are a very poor proxy for your actual impact.  * Some other reason that

I think your arguments do suggest good reasons why nuclear risk might be prioritized lower; since we operate on the most effective margin, as you note, it is also possible at the same time for there to be significant funding margins in nuclear that are highly effective in expectation.

2
Nathan Young
8mo
Do you work on researching nuclear risk? How do you think this disagreement could be more usefully delineated. It seems like there is some interesting disagreement here?

My point is precisely that you should not assume any view. My position is that the uncertainties here are significant enough to warrant some attention to nuclear war as a potential extinction risk, rather than to simply bat away these concerns on first principles and questionable empirics.

Where extinction risk is concerned, it is potentially very costly to conclude on little evidence that something is not an extinction risk. We do need to prioritize, so I would not for instance propose treating bad zoning laws as an X-risk simply because we can't demonstra... (read more)

1
Nathan Young
8mo
My argument does say something about how nuclear risk shoud be prioritised. It is a lower priority if both existed. Maybe much lower. The complicated thing is that nuclear risks do exist whereas biorisk and AI risk are much more speculative in terms of actually existing. In this sense I can believe nuclear should be funded more.

If you leave 1,000 - 10,000 humans alive, the longterm future is probably fine

This is a very common claim that I think needs to be defended somewhat more robustly instead of simply assumed. If we have one strength as a community, is in not simply assuming things.

My read is that the evidence here is quite limited, the outside view suggests that losing 99.9999% of a species / having a very small population is a significant extinction risk, and that the uncertainty around the long-term viability of collapse scenarios is enough reason to want to avoid near-extinction events.

2
Nathan Young
8mo
Why do I think 1,000 -10,000 humans is probably (60 - 90%) fine? According to Luisa Rodriguez, you need about 300 people to rebuild the human race. These people seem likely to be very incentivised towards survival - humans generally like surviving. It would be awful for them, sure, but the question is would they rebuild us as a species?  And I think the answer is probably.  And let's remember that this is the absolute worst case scenario. The human race has twice dropped nuclear bombs and then never again. It seems a big leap to imagine that not only will we do so but we will wipe ourselves out to the extent of only 1 such group.  Every successive group that could rebuild the human race is extra. I imagine that actually 100s of millions would survive an actual worldwide nuclear war, so the point we are litigating is a very small chance anwyay.  I don't really know what base rates I'd use here. Feels like you want natural disasters rather than predation. When the meteor hit do we know how population size affected repopulation? Even then, humans are just way more competent than any other animals. So as I said originally we might be looking at a 10 - 40% chance given the near worst case scenario, but I don't buy your outside view.  I'd be curious what others outside views are here and if anyone has actual base rates on disaster driven animal populations and repopulation.    As an aside,  I disagree. I've said what I think, you can push back on it if you want, but why is it bad to "simply assume" my view rather than yours? 

Has there been any formal probabilistic risk assessment on AI X-risk? e.g. fault tree analysis or event tree analysis — anything of that sort?

9
aogara
9mo
Here’s a fault tree analysis: https://arxiv.org/abs/2306.06924 Review of risk assessment techniques that could be used: https://arxiv.org/abs/2307.08823 Applying ideas from systems safety to AI: https://arxiv.org/abs/2206.05862 Applying ideas from systems safety to AI (part 2): https://arxiv.org/abs/2302.02972 Applying AI to ideas from systems safety (lol): https://arxiv.org/abs/2304.01246
3
Erich_Grunewald
9mo
I recently learned of this effort to model AI x-risk, which may be similar to the sort of thing you're looking for, though I don't think they actually put numbers on the parameters in their model, and they don't use any well-known formal method. Otherwise I suppose the closest thing is the Carlsmith report, which is a probabilistic risk assessment, but again not using any formal method.

I disagree with the valence of the comment, but think it reflects legitimate concerns.

I am not worried that "HLI's institutional agenda corrupts its ability to conduct fair-minded and even-handed assessment." I agree that there are some ways that HLI's pro-SWB-measurement stance can bleed into overly optimistic analytic choices, but we are not simply taking analyses by our research partners on faith and I hope no one else is either. Indeed, the very reason HLI's mistakes are obvious is that they have been transparent and responsive to criticism.

We disagree... (read more)

I agree that there are some ways that HLI's pro-SWB-measurement stance can bleed into overly optimistic analytic choices, but we are not simply taking analyses by our research partners on faith and I hope no one else is either. 

Individual donors are, however, more likely to take a charity recommender's analysis largely on faith -- because they do not have the time or the specialized knowledge and skills necessary to kick the tires. For those donors, the main point of consulting a charity recommender is to delegate the tire-kicking duties to someone who has the time, knowledge, and skills to do that.

I guess I would very slightly adjust my sense of HLI, but I wouldn't really think of this as an "error." I don't significantly adjust my view of GiveWell when they delist a charity based on new information.

I think if the RCT downgrades StrongMinds' work by a big factor, that won't really introduce new information about HLI's methodology/expertise. If you think there are methodological weaknesses that would cause them to overstate StrongMinds' impact, those weaknesses should be visible now, irrespective of the RCT results.

2
Nathan Young
10mo
So, for clarity, you disagree with @Gregory Lewis[1] here: 1. ^ How do i do the @ search?

I can also vouch for HLI. Per John Salter's comment, I may also have been a little sus early (sorry Michael) on but HLI's work has been extremely valuable for our own methodology improvements at Founders Pledge. The whole team is great, and I will second John's comment to the effect that Joel's expertise is really rare and that HLI seems to be the right home for it.

0
MichaelPlant
10mo
Hello Matt and thanks for your overall vote of confidence, including your comments below to Nathan.  Could you expand on what you said here? I'm curious to know why you were originally suspicious and what changed your mind. Sorry if you've already stated that below. 
5
Nathan Young
10mo
I appreciate this kind of transparent vouching for orgs. Makes it easier to discuss what's going on.  How do you think you'll square this if the forthcoming RCT downgrades StrongMind's work by a factor of 4 or more? I'm confused about how HLI could miss this error (if it happens) That said, as John says their actual produced work could still be very cheap at this price.

Just a note here as the author of that lobbying post you cite: the CEA including the 2.5% change in chance of success is intended to be illustrative — well, conservative, but it's based on nothing more than a rough sense of effect magnitude from having read all those studies for the lit review. The specific figures included in the CEA are very rough.  As Stephen Clare pointed out in the comments, it's also probably not realistic to have modeled that is normal on the [0,5] 95% CI.

2
Ben Stewart
10mo
Thanks for the clarification!

Hey Vasco, you make lots of good points here that are worth considering at length. These are topics we've discussed on and off in a fairly unstructured way on the research team at FP, and I'm afraid I'm not sure what's next when it comes to tackling them. We don't currently have a researcher dedicated to animal welfare, and our recommendations in that space have historically come from partner orgs.

Just as context, the reason for this is that FP has historically separated our recommendations into three "worldviews" (longtermism, current generations, and ani... (read more)

2
Vasco Grilo
1y
Thanks for sharing your thought, Matt!

Hey Matthew, thanks for sharing this. Can you provide some more information (or link to your thoughts elsewhere) on why fervor around UV-C is misplaced? As you know, ASHRAE Standards 185.1 and 185.2 concern testing of UV devices for germicidal irradiation, so I'd be particularly interested to know if this was an area that ASHRAE itself had concluded was unpromising.

3
MatthewDahlhausen
1y
Here's a related comment from last year: https://forum.effectivealtruism.org/posts/nopFhTtoiyGX8Bs7G/uvc-air-purifier-design-and-testing-strategy?commentId=ZywtAzPB2Ci5PLfPC UV systems have been around for ~100 years. They work great in some specific applications. Newer UV-C technology is a marginal improvement, but doesn't significantly address the cost, design expertise, and maintenance challenges that have kept UV systems from widespread use. Air filters are generally better for most applications. I do expect we will see more UV-C systems in particular applications, but it is far from the one-technology-to-rule-them-all solution that the EA community seems to think it is. This follows a historic pattern of the EA community generally over-hyping singular technology solutions to major problems in other cause areas, probably because of the techno-optimist worldview that many EAs have. See ASHRAE's guidance on filtration and air cleaning technologies for more details and comparisons: https://www.ashrae.org/technical-resources/filtration-disinfection

I thought of some other down-the-line feature requests

  • Google Sheets integration (we currently already store our forecasts in a Google sheet)
  • Relatedly, ability to export to CSV (does this already exist and I just missed it?)
  • Ability to designate a particular resolver
  • Different formal resolution mechanisms, like a poll of users.

Ah, great! I think it would be nice to offer different aggregation options, though if you do offer one I agree that geo mean of odds is the best default. But I can imagine people wanting to use medians or averages, or even specifying their own aggregation functions. Especially if you are trying to encourage uptake by less technical organizations, it seems important to offer at least one option that is more legible to less numerate people.

I have already installed this and started using this at Founders Pledge. Thanks for making this! I've been wanting something like this for a long time.

Some feature requests:

  • Aggregation choices (e.g. geo mean of odds would be nice)
  • Brier scores for users
  • Calibration curves for users
5
Adam Binks
1y
I've added a basic calibration curve, thanks for the suggestion!  You can find it in the app's Home tab (click on Fatebook in the left sidebar > Home tab at the top) once at least one question you've forecasted on has resolved.
3
Adam Binks
1y
Great, glad to hear it! Geo mean of odds is a good idea - it's probably a more sensible default. How would you feel about us using that everywhere, instead of the current arithmetic mean? You can see your own absolute and relative Brier score in the app home (click Fatebook in the sidebar). If you're thinking of a team-wide leaderboard - that's on our list! Though some users said they wouldn't like this to avoid Goodharting, so I've not prioritised it so far, and will include a team-wide toggle if we add it. We'll add this soon!

Honestly, what surprises me most here is how similar all four organizations' numbers are across most of the items involved

 

This was also gratifying for us to see, but it's probably important to note that our approach incorporates weights from both GiveWell and HLI at different points, so the estimates are not completely independent.

Thanks, bruce — this is a great point. I'm not sure if we would account for the costs in the exact way I think you have done here, but we will definitely include this consideration in our calculation.

I haven't thought extensively  about what kind of effect size I'd expect, but I think I'm roughly 65-70% confident that the RCT will return evidence of a detectable effect.

But my uncertainty is more in terms of rating upon re-evaluating the whole thing. Since I reviewed SM last year, we've started to be a lot more punctilious about incorporating various discounts and forecasts into CEAs. So on the one hand I'd naturally expect us to apply more of those discounts on reviewing this case, but on the other hand my original reason for not discounting HLI's... (read more)

As promised, I am returning here with some more detail. I will break this (very long) comment into sections for the sake of clarity.

My overview of this discussion

It seems clear to me that what is going on here is that there are conflicting interpretations of the evidence on StrongMinds' effectiveness. In particular, the key question here is what our estimate of the effect size of SM's programs should be. There are other uncertainties and disagreements, but in my view, this is the essential crux of the conversation. I will give my own (personal) interpretat... (read more)

During the re-evaluation, it would be great if FP could also check the partnership programme by StrongMinds - e.g. whether this is an additional source of revenue for them, and what the operational costs of the partners who help treat additional patients for them are. At the moment these costs are not incorporated into HLI's CEA, but partners were responsible for ~50 and ~80% of the clients treated in 2021 and 2022 respectively. For example, if we crudely assume costs of treatment per client are constant regardless of whether it's treated by StrongMinds or... (read more)

4
Simon_M
1y
Out of interest what do your probabilities correspond to in terms of the outcome from the Ozler RCT? (Or is your uncertainty more in terms of what you might find when re-evaluating the entire framwork?)

Hey Simon, I remain slightly confused about this element of the conversation. I take you to mean that, since we base our assessment mostly on HLI's work, and since we draw different conclusions from HLI's work than you think are reasonable, we should reassess StrongMinds on that basis. Is that right?

If so, I do look forward to your thoughts on the HLI analysis, but in the meantime I'd be curious to get a sense of your personal levels of confidence here — what does a distribution of your beliefs over cost-effectiveness for StrongMinds look like?

8
Simon_M
1y
I'm not sure exactly what you've done, so it's hard for me to comment precisely. I'm just struggling to see how you can be confident in a "6x as effective as GD" conclusion. So there are two sides to this: 1. Is my confidence in HLI's philisophical views. I have both spoken to Joel and read all their materials several times and thinkI understand their views. I am sure I do not fully agree with them and I'm not sure how much I believe them. I'd put myself at roughly 30% that I agree with their general philosophy. This is important because how cost-effective you believe StrongMinds are is quite sensitive to philisophical assumptions. (I plan on expanding upon this when discussing HLI) 2. Under HLI's philosophical assumptions, I think I'm roughly speaking: 10% SM is 4-8x as good at GiveDirectly 25% SM is 1-4x as good as GiveDirectly 35% SM is 0.5-1x as good as GiveDirectly 30% SM not effective at all So roughly speaking under HLI's assumptions I think StrongMinds is roughly as good as GiveDirectly. I think you will probably say on this basis that you'd still be recommending StrongMinds based on your risk-neutral principle but I think this underestimates quite how uncertain I would expect people to be in the HLI worldview. (I also disagree with being risk-neutral, but I suspect that's a discussion for another day!)

Fair enough. I think one important thing to highlight here is that though the details of our analysis have changed since 2019, the broad strokes haven’t — that is to say, the evidence is largely the same and the transformation used (DALY vs WELLBY), for instance, is not super consequential for the rating.

The situation is one, as you say, of GIGO (though we think the input is not garbage) and the main material question is about the estimated effect size. We rely on HLI’s estimate, the methodology for which is public.

I think your (2) is not totally fair to S... (read more)

“I think my main takeaway is my first one here. GWWC shouldn't be using your recommendations to label things top charities. Would you disagree with that?”

Yes, I think so- I’m not sure why this should be the case. Different evaluators have different standards of evidence, and GWWC is using ours for this particular recommendation. They reviewed our reasoning and (I gather) were satisfied. As someone else said in the comments, the right reference class here is probably deworming— “big if true.”

The message on the report says that some details have changed, bu... (read more)

Yes, I think so- I’m not sure why this should be the case. Different evaluators have different standards of evidence, and GWWC is using ours for this particular recommendation. They reviewed our reasoning and (I gather) were satisfied. As someone else said in the comments, the right reference class here is probably deworming— “big if true.”

I'm afraid that doesn't make me super impressed with GWWC, and it's not easy for non-public reasoning to be debunked. Hopefully you'll publish it and we can see where we disagree.

I think there's a big difference between ... (read more)

Hi Simon, thanks for writing this! I’m research director at FP, and have a few bullets to comment here in response, but overall just want to indicate that this post is very valuable. I’m also commenting on my phone and don’t have access to my computer at the moment, but can participate in this conversation more energetically (and provide more detail) when I’m back at work next week.

  • I basically agree with what I take to be your topline finding here, which is that more data is needed before we can arrive at GiveWell-tier levels of confidence about Strong

... (read more)
8
Tanae
1y
Is there any way we can get more details on this? I recently made a blogpost using Bayesian updates to correct for post-decision surprise in GiveWell's estimates, which led to a change in the ranking of New Incentives from 2nd to last in terms of cost effectiveness among Top Charities. I'd imagine (though I haven't read the studies) that the uncertainty in the Strong Minds CEA is / should be much larger.  For that reason, I would have guessed that Strong Minds would not fare well post-Bayesian adjustment, but it's possible you just used a different (reasonable) prior than I did, or there is some other consideration I'm missing? Also, even risk neutral evaluators really should be using Bayesian updates (formally or informally) in order to correct for post-decision surprise. (I don't think you necessarily disagree with me on this, but it's worth emphasizing that valuing GW-tier levels of confidence doesn't imply risk aversion.)
7
ClimateDoc
1y
"we estimate StrongMinds at roughly 6x GD" - this seems to be about 2/3 what HLI estimate the relative impact to be (https://forum.effectivealtruism.org/posts/zCD98wpPt3km8aRGo/happiness-for-the-whole-household-accounting-for-household) - it's not obvious to me how and why your estimates differ - are you able to say what is the reason for the difference? (Edited to update to a more recent analysis by HLI)

The 2019 report you link (and the associated CEA) is deprecated— FP hasn’t been resourced to update public-facing materials, a situation that is now changing—but the proviso at the top of the page is accurate: we stand by our recommendation.

The page doesn't say deprecated and GWWC are still linking to it and recommending it as a top charity. I do think your statements here should be enough for GWWC to remove them as a top charity. 

This is what triggered the whole thing in the first place - I have had doubts about StrongMinds for a long time (I private... (read more)

Thanks for this! Useful to get some insight into the FP thought process here.

The effect sizes observed are very large, but it’s important to place in the context of StrongMinds’ work with severely traumatized populations. Incoming PHQ-9 scores are very, very high, so I think  ...  2) I’m not sure that our general priors about the low effectiveness of therapeutic interventions are likely to be well-calibrated here.

(emphasis added)
 

Minor nitpick (I haven't personally read FP's analysis / work on this):
Appendix C (pg 31) details the recruitment... (read more)

Hey Nick, thanks for this very valuable  experience-informed comment. I'm curious what you make of the original 2002 RCT that first tested IPT-G  in Uganda. When we (at Founders Pledge) looked at StrongMinds (which we currently recommend, in large part on the back of HLI's research), I was surprised to see that the results from the original RCT lined up closely with the pre/post scores reported by recent program participants.

Would your take on this result be that participants in the treated group were still basically giving  what they saw as... (read more)

2
NickLaing
1y
Matt these are fantastic questions that I definitely don't have great answers to, but here are a few thoughts. First I'm not saying at all that the Strong minds intervention is likely useless - I think it is likely very useful. Just that the positive effects may well be grossly overstated for the reasons outlined above. My take on the result of that original 2002 RCT and Strong Minds. Yes like you say in both cases it could well be that the treatment group are giving positive answers both to appease the interviewer (Incredibly the before and after interviews were done by the same researcher in that study which is deeply problematic!)  and because they may have been hoping positive responses might provide them with further future help. Also in most of these studies, participants are given something physical for being part of the intervention groups. Perhaps small allowances for completing interviews, or tea and biscuits during the sessions. These tiny physical incentives can be more appreciated than the actual intervention. Knowing World  Vision this would almost certainly be the case I have an immense mistrust of World vision for a whole range of reasons, who were heavily involved in that famous 2002 RCT. This is due to their misleading advertising and a number of shocking experiences of their work here in Northern Uganda which I won't expand on. I even wrote a blog about this a few years ago, encouraging people not to give them money.  I know this may be a poor reason to mistrust a study but my previous experience heavily biases me all the same. Great point about the NBER paper which featured a pure control group. First it was a different intervention - individual CBT not group therapy.  Second it feels like the Kenyan study was more dispassionate than some of the other big ones. I might be wrong but a bunch of the other RCTs are partly led and operated by organisations with something to prove. I did like that the Kenyan RCT felt less likely to be biased as t

I agreed with your comment (I found it convincing) but downvoted it because if I was a first-time poster here, I would be much less likely to post again after having my first post characterized as foolish.

As one of many “naive functionalists”, I found the OP was very valuable as a challenge to my thinking, and so I want to come down strongly against discouraging such posts in any way.

4
Jackson Wagner
2y
Ah, I didn't notice that this was a totally new post! Honestly the writing style and polish felt like such a good match for the Forum, I assumed it must be coming from somebody who posts here all the time. In this new context -- sorry, Marcusarvan, for immediately jumping onto a disagreement and not even noticing all the common ground evident via your writing style and EA interests!

I agree- the EA community claims to be "open to criticism" but having someone comment that a post is foolish on a first time poster's well articulated and argued post is quite frankly really disappointing. 

In addition, the poster is a professional and has valuable knowledge regardless of how you feel about the merits of their argument. 

I'm a university student and run an EA group at my university. I really wish the community would be more open to professionals like this poster who aren't affiliated with an EA organization, but can contribute different perspectives that aren't as common within the community. 

 

These seem like broadly reasonable heuristics, but they kick the can on who is an expert, which is where most of the challenge in deference lies.

The canonical (recent) example of this is COVID, when doctors and epidemiologists, who were perceived by the general public as the relevant experts, weighed in on questions of public policy, in many cases giving the impression of consensus in their communities. I think there is a good argument to be made that public policy “experts” were in fact better-placed to give recommendations in many of these issues. Regard... (read more)

1
Sharmake
2y
True, it can be a bit of a challenge to use this heuristic. Though do note my 3rd caveat, it isn't a replacement for EV calculations, and a lot of high impact areas have this issue.

(I am research director at FP)

Thanks for all of your work on this analysis, Vasco. We appreciate your thoroughness and your willingness to engage with us beforehand. The work is obviously methodologically sound and, as Johannes indicated, we generally agree that climate is not among the top bets for reducing existential risk.

I think that "mitigating existential risk as cost-effectively as possible" is entailed by the goal of doing as much good as possible in the world, which is why FP exists. To be absolutely clear, FP's goal is to do the maximum possible ... (read more)

3
Vasco Grilo
2y
Thanks, Matt! I think the above is a very valuable explanation of FP's mission, and have edited the last 2 paragraphs of the Introduction to point to your and Johannes' comments.

Do you have any plans for interoperability with other PPLs or languages for statistical computing? It would be pretty useful to be able to, e.g. write a model in Squiggle and port it easily to R or to PyMC3, particularly if Bayesian updating is not currently supported in Squiggle. I can easily imagine a workflow where we use Squiggle to develop a prior, which we'd then want to update using microdata in, say, Stan (via R).

9
NunoSempere
2y
I think that for the particular case where Squiggle produces a distribution (as opposed to a function that produces a distribution), this is/should be possible.
3
Ozzie Gooen
2y
No current plans. I think it would be tricky, because Squiggle both supports some features that other PPL's don't, and because some of them require stating information about variables upfront, which Squiggle doesn't. Maybe it's possible for subsets of Squiggle code. Would be happy to see experimentation here. I think one good workflow would be to go the other way; use a PPL to generate certain outcomes, then cache/encode these in Squiggle for interactive use. I described this a bit in this sequence. https://www.lesswrong.com/s/rDe8QE5NvXcZYzgZ3

I very strongly downvoted this comment because I think that personal attacks of any sort have a disproportionately negative impact on the quality of discussion overall, and because responding to a commenter's identity or background instead of the content of their comment is a bad norm.

-6
Dem0sthenes
2y

Founders Pledge is hiring an Applied Researcher to work with our climate lead evaluating funding opportunities, finding new areas to research within climate, evaluating different theories of change, and granting from FP's Climate Fund.

We're open to multiple levels of seniority, from junior researchers all the way up to experienced climate grantmakers. Experience in climate and a familiarity with energy systems is a big plus, but not 100% necessary.

Our job listing is here. Please note that the first round consists of a resume screen and a preliminary task. ... (read more)

Something I've considered making myself is a Slackbot for group decision-making: forecasting, quadratic voting, etc. This seems like it would be very useful for lots of organizations and quite a low lift. It's not the kind of thing that seems easily monetizable at first, but it seems reasonable to expect that if it provides valuable, it could be the kind of thing that people would eventually have to buy "seats" for in larger organizations.

I appreciate your taking the time to write out this idea and the careful thought that went into your post. I liked that it was kind of in the form of a pitch, in keeping with your journalistic theme. I agree that EAs should be thinking more seriously about journalism (in the broadest possible sense) and I think that this is as good a place as any to start. I want to (a) nitpick a few things in your post with an eye to facilitating this broader conversation and (b) point out what I see as an important potential failure mode for an effort like this.

You chara... (read more)

1
Ben Williamson
2y
This is great! I appreciate the detailed response and the friendly but valuable critique. As much as anything, I wanted to start a conversation around how a project like this may work best and these are some really useful points. I think you're right that there are competing goals in what I've laid out and also that a project reporting from an EA lens on general topics might be more impactful in the long term than one focused on reporting on EA topics.
2
Chris Leong
2y
I agree that a trade publication could potentially be valuable - my main issue would be that attempts at outreach to the existing philanthropic community haven't been very successful before. Nonetheless, I suspect that if it were able to produce sufficiently high-quality content - including content about topics outside of the traditional EA areas, I suspect I would gain readers and influence.

While I’m skeptical about the idea that particular causes you’ve mentioned could truly end up being cost effective paths to reducing suffering, I’m sympathetic to the idea that improving the effectiveness of activity in putatively non-effective causes is potentially itself effective. What interventions do you have in mind to improve effectiveness within these domains?

3
freedomandutility
3y
I think the interventions would be very specific to the domain. I mentioned an intervention to direct pro-Palestinian activism towards a tangible goal, and with redirecting western anti-racism work towards international genocide prevention, this could possibly be done by getting western anti-racism organisations to partner with similar organisations in countries with greater risk of genocides, which could lead to resource / expertise sharing over a long period of time.

Now that you’ve given examples, can you provide an account of how increased funding in these areas can lead to improved well-being / preserves lives or DALYs / etc in expectation? Do you expect that targeted funds could be cost-competitive with GW top charities or likewise?

1
freedomandutility
3y
So in both of the examples provided, EAs would be funding / carrying out interventions that improve the effectiveness of other work, and it is this other work that would improve well-being / preserve lives in expectation. Because I suspect that these interventions would be relatively cheap, and because this other work would already have lots of resources behind it, I think these interventions would slightly improve the effectiveness with which a large amount of resources are spent, to the extent that the interventions could compare with GW top charities in terms of expected value.

To clarify, I'm not sure this is likely to be the best use of any individual EA's time, but I think it can still be true that it's potentially a good use of community resources, if intelligently directed.

I agree that perhaps "constitutionally" is too strong - what I mean is that EAs tend (generally) to have an interest in / awareness of these broadly meta-scientific topics.

In general, the argument I would make would be for greater attention to the possibility that mainstream causes deserve attention and more meta-level arguments for this case (like your post).

Thanks for this! It seems like much of the work that went into your CEA could be repurposed for explorations of other potentially growth- or governance-enhancing interventions. Since finding such an intervention would be quite high-value, and since the parameters in your CEA are quite uncertain, it seems like the value of information with respect to clarifying these parameters (and therefore the final ROI distribution) is probably very high.

Do you have a sense of what kind of research or data would help you narrow the uncertainty in the parameter inputs of your cost-effectiveness model?

7
David Rhys Bernard
3y
I'm not convinced that our CEA is particularly useful for more generalised interventions. All we really do is assume that the intervention causes some growth increase (a distribution rather than a point estimate) and then model expected income with the intervention, with the intervention 10 years later and with no intervention. The amount the intervention increases growth is the key parameter and is very uncertain so further research on this will have the highest VoI, but this will be different for each intervention. We treat how the intervention increases growth as a black box so I think looking inside the box and trying to understand the mechanisms better would shed some light on how robust the assumed growth increase is and how we might expect it to generalise to other contexts. Furthermore, we only model the direct benefits of the growth intervention. In general, I'd expect the indirect effects to be larger and our modelling approach doesn't say anything about these so I expect looking into these indirect benefits, perhaps via an alternative model, to have higher VoI than further modelling of the direct benefits. For charter cities in particular, we could probably further tighten the bounds on the direct benefits by getting more rigorous information on city population growth rates and the correlation between population growth and income growth.

On the face of it, it seems like researching and writing about "mainstream" topics is net positive value for EAs for the reasons you describe, although not obviously an optimal use of time relative to other competing opportunities for EAs. I've tried to work out in broad strokes how effective it might be to move money within putatively less-effective causes, and it seems to me like (for instance) the right research, done by the right person or group, really could make a meaningful difference in one of these areas.

Items 2.2 and 2.3 (in your summary) are, to... (read more)

2
MichaelA
3y
Thanks for this comment. I think I essentially agree with all your specific points, though I get the impression that you're more optimistic about trying to get "better answers to mainstream questions" often being the best use of an EA's time. That said:  * this is just based mainly on something like a "vibe" from your comment (not specific statements) * my own views are fairly tentative anyway * mostly I think people also need to consider specifics of their situation, rather than strongly assuming either that it's pretty much always a good idea to try to get "better answers" on mainstream questions or that it's pretty much never a good idea to try that One minor thing I'd push back on is "especially for EAs, who are constitutionally hyper-aware of the pitfalls of bad research, have high standards of rigor, and are often quantitatively sophisticated." I think these things are true on average, but "constitutionally" is a bit too strong, and there is also a fair amount of bad research by EAs, low standards of rigour among EAs, and other problems. And I think it's importnat that we remember that (though not in an over-the-top or self-flagellating way, and not with a sort of false modesty that would guide our behaviour poorly).

I think about this all the time. It seems like a really high-value thing to do not just for the sake of other communities but even from a strictly EA perspective— discourse norms seem to have a real impact on the outcome of decision-relevant conversations, and I have an (as-yet unjustified) sense that EA-style norms lead to better normative outcomes. I haven't tried it, but I do have a few isolated, perhaps obvious observations.

  • For me at least, it is easier to hew to EA discussion norms when they are, in fact, accepted norms. That is, assuming the best int
... (read more)

I guess a more useful way to think about this for prospective funders is to move things about again. Given that you can exert c/x leverage over funds within Cause Y, then you're justified in spending c to do so provided you can find some Cause Y such that the distribution of DALYs per dollar meets the condition...

...which makes for a potentially nice rule of thumb. When assessing some Cause Y, you need only ("only") identify a plausibly best or close-to-best opportunity, as well as the median one, and work from there.

Obviously this condition... (read more)

Under what circumstances is it potentially cost-effective to move money within low-impact causes?

This is preliminary and most likely somehow wrong.  I'd love for someone to have a look at my math and tell me if (how?) I'm on the absolute wrong track here.

Start from the assumption that there is some amount of charitable funding that is resolutely non-cause-neutral. It is dedicated to some cause area Y and cannot be budged. I'll assume for these purposes that DALYs saved per dollar is distributed log-normally within Cause Y:

I want t... (read more)

4
NunoSempere
3y
Interesting. You might get more comments as a top-level post.
1
Matt_Lerner
3y
I guess a more useful way to think about this for prospective funders is to move things about again. Given that you can exert c/x leverage over funds within Cause Y, then you're justified in spending c to do so provided you can find some Cause Y such that the distribution of DALYs per dollar meets the condition... Q0.9≥c100x+Q0.5 ...which makes for a potentially nice rule of thumb. When assessing some Cause Y, you need only ("only") identify a plausibly best or close-to-best opportunity, as well as the median one, and work from there. Obviously this condition holds for any distribution and any set of quintiles, but the worked example above only indicates to me that it's a plausible condition for the log-normal.

What do you see as the consequentialist value of doing journalism? What are the ways in which journalists can improve the world? And do you believe these potential improvements are measurable?

3
Tom Chivers
3y
I think it would be very hard to have a functioning democratic state without journalism of some kind. I may be overstating my industry's importance, but if you don't know what the government is doing, or how the machinery of the state is operating, or what the lives of the citizens are like, then how can you participate in your democracy? And you can't rely on the government to tell you. So even though most journalism is not vital to democracy, if there was no journalism, there would be very little to stop the government from behaving how it liked. I also think that in my field of journalism, science writing, there's a lot of value in translating abstruse-but-important research and thinking into readable or even enjoyable material for laypeople. Also, you can convince people of things that are true, and help people make good decisions (I hope). Plus, things like criticism are helpful for readers in allowing them to find the books/films/theatre they might enjoy (and I think they have a value even given the existence of crowdsourced review sites like Rotten Tomatoes). And, of course, people enjoy reading/watching, and that is a good in itself.  For someone like me who has no other skills than interviewing clever people and writing down what they say, I suspect that journalism is one of the places I could do the most good, because I can do it well. Of course, all the good outcomes I just mentioned are reliant on the journalist in question being good at their job, but that's probably true of all careers, isn't it? On measuring it: re democracy at least, I guess you could try to do some sort of study looking at countries with strong independent journalism vs those without, but they would be so horribly confounded I doubt you could get good numbers on it. 

One thing to note here is that lots of commonly-used power law distributions have positive support. Political choices can and sometimes do have dramatically negative effects, and many of the catastrophes that EAs are concerned with are plausibly the result of those choices (like nuclear catastrophe, for instance). 

So a distribution that describes the outcomes of political choices should probably have support on the whole real line, and you wouldn't want to model choices with most simple power-law distributions.  But you might be on to something--... (read more)

4
Daniel_Eth
3y
Worth noting that if some political choices have very large negative outcomes, then choosing political paths that avoid those outcomes would have very positive counterfactual impact, even if no one sees it.
1
felix.h
3y
Good point! I can't say I have an immediate response, but I'm gonna think a bit more about this.

I read this post with a lot of interest; it has started to seem more likely to me lately that spreading productive, resilient norms about decision-making and altruism is a more effective means of improving decisions in the long run than any set of particular institutional structures. The knock-on effects of such a phenomenon would, on a long time scale, seem to dwarf the effects of many other ostensibly effective interventions.

So I get excited about this idea. It seems promising.

But some reflection about what is commonly considered precedent for something ... (read more)

2
Owen Cotton-Barratt
3y
I agree with this. I don't think science has the attractor property I was discussing, but it has this other attraction of being visibly useful (which is even better). I was trying to use science as an example of the self-correction mechanism. Yes, this is the sense of self-propagating that I intended.

I very strongly upvoted this because I think it's highly likely to produce efficiencies in conversation on the Forum, to serve as a valuable reference for newcomers to EA, and to act as a catalyst for ongoing conversation.

I would be keen to see this list take on life outside the forum as a standalone website or heavily moderated wiki, or as a page under CEA or somesuch, or at QURI.

I feel it should be pointed out that there already is a similar standalone wiki causeprioritization.org and until recently there was another similar website PriorityWiki but I think that neither of them have received much traffic.

7
NunoSempere
3y
Thanks!  Ozzie has been proposing something like that. Initially, an airtable could be nice for visualization. 

I'm not sure why this is being downvoted. I don't really have an opinion on this, but it seems at least worth discussing. OP, I think this is an interesting idea.

John Lewis Gaddis' The Cold War: A New History contains a number of useful segments about the nuclear tensions between the U.S. and the U.S.S.R., insightful descriptions of policymakers' thinking during these moments, and a consideration of counterfactual histories in which nuclear weapons might have been deployed. I found it pretty useful in terms of helping me get a picture of what decision-making looks like when the wrong decision means (potentially) the end of civilization.

How harmful is a fragmented resume? People seem to believe this isn't much of a problem for early-career professionals, but I'm 30, and my longest tenure was for two and a half years (recently shorter). I like to leave for new and interesting opportunities when I find them, but I'm starting to wonder whether I should avoid good opportunities for the sake of appearing more reliable as a potential employee.

I think it depends a lot on industry. In the world of startups frequently changing jobs doesn't seem that unusual at all. In finance, on the other hand, I would be very suspicious of someone who moved from one hedge fund to another every two years.

It also depends a bit on the role. A recent graduate who joins an investment bank as an analyst is basically expected to leave after two years; but if a Director leaves after two years that is a sign that something was wrong. Working as a teacher for two years and then quitting looks bad, unless it was Teach for America, in which case it is perfectly normal.

8
Benjamin_Todd
3y
Hi Matt, This is a common concern, though I think it's helpful to zoom out a bit – what employers most care about is that you can prove to them that you can solve the problems they want solved. Insofar as that relates to your past experience (which is only one factor among many they'll look at), my impression[1] is that what matters is whether you can tell a good story about (i) how your past experience is relevant to their job and (ii) what steps have let you to wanting to work for them. This is partly an exercise in communication. If your CV doesn't naturally lead to the job, you might want to spend more time talking with friends / advisors about how to connect your past experience to what they're looking for. It depends even more on whether you had good reasons for changing, and whether you've built relevant career capital despite it. I can't evaluate the latter from here, so I might throw the question back to you: do you think you've changed too often, or was each decision good? I'm sympathetic to the idea that early career, a rule of thumb for exploration like "win, stick; lose, shift" makes sense (i.e. if a career path is going ahead of expectations, stick with it; and otherwise shift), and that can lead to lots of shifting early on if you get unlucky. However, you also need to balance that with staying long enough to learn skills and get achievements, which increase your career capital. ---------------------------------------- 1. *How to successfully apply to jobs isn't one of my areas of expertise, though I have experience as an employer and in marketing, and have read about it some. ↩︎

First, congratulations. This is impressive, you should be very proud of yourself, and I hope this is the beginning of a long and fruitful data science career (or avocation) for you.
 

What is going on here?


I think the simplest explanation is that your model fit better because you trained on more data. You write that your best score was obtained by applying XGBoost to the entire feature matrix, without splitting it into train/test sets. So assuming the other teams did things the standard way, you were working with 25%-40% more data to fit the model. In a... (read more)

5
Tsunayoshi
3y
Are you sure that this is the standard way in competitions? It is absolutely correct that before the final submission, one would find the best model by fitting it on a train set and evaluating it on the test set. However, once you found a best performing model that way, there is no reason not to train the model with the best parameters on the train+test set, and submit that one. (Submission are the predictions of the model on the validation set, not the parameters of the model).  After all, more data equals better performance.  

I wonder if the forum  shouldn't encourage a class of post (basically like this one) that's something like "are there effective giving opportunities in X context?" Although EA is cause-neutral, there's no reason why members shouldn't take the opportunity provided by serendipity to investigate highly specific scenarios and model "virtuous EA behavior." This could be a way of making the forum friendlier to visitors like the OP, and a way for comments to introduce visitors to EA concepts in a way that's emotionally relevant.

1
ld25
4y
If I understand you correctly, I agree. I understand the reason for quoting GiveWell's framework, however, I think that it is potentially discouraging to someone who is trying to do the most good in a context that they care about. That's not to say that nobody should ever say 'maybe there are more neglected causes that you may not have thought about', but the EA community certainly shouldn't be giving the impression that we follow some strict ideology that no-one can challenge.
Load more