As far as I can tell The Long-Term Future Fund (LTFF) wants to fund work that will influence public policy. They say they are looking to fund “policy analysis, advocacy ...” work and post in policy Facebook groups asking for applications.
However, as far as I can tell, in practice, the LTFF appears to never (or only very rarely) fund such projects that apply for funding. Especially new projects.
My view that such funding is rare is based on the following pieces of evidence:
- Very few policy grants have been made. Looking at the payout reports of the LTFF, they funded a grant for policy influencing work in 2019 (to me, for a project that was not a new project). Upon asking them they say they have funded at least one more policy grant that has not been written up.
- Very many policy people have applied for LTFF grants. Without actively looking for it I know of: someone in an established think tank looking for funds for EA work, 3 groups wanting to start EA think tank type projects, a group wanting to do mass campaigning work. All these groups looked competent and were rejected. I am sure many others apply too. I know one of these has gone on to get funding elsewhere (FTX).
- Comments from staff at leading EA orgs. In January last year, a senior staff member at a leading EA institution mentioned, to my surprise, that EA funders tend not to fund any new longtermist policy projects (except perhaps with very senior trusted people like OpenPhil funding CSET). Recently I spoke to someone at CEA about this and asked if it matched their views too and they said they do think there is a problem here. Note this was about EA funders in general, not specifically the LTFF.
- Comments from EA policy folk looking for funding. There seems to be (at least there was in 2021) a general view from EAs working in the policy space that it has been very hard to find funding for policy work. Note this is about EA funders in general, not the LTFF.
- Odd lines of questioning. When I applied, the line of questioning was very odd. I faced an hour of: Why should we do any policy stuff? Isn't all policy work a waste of time? Didn’t [random unconnected policy thing] not work? Etc. Of course it can be useful to check applicants have a good understanding of what they are doing, but it made me question whether they wanted to fund policy work at all. [Edit: apparently they do this style of questioning for various topics like AI safety research not just policy, so maybe not an issue.]
- Odd feedback. Multiple applicants to the LTFF have reported receiving feedback along the lines of: We see high downside risks but cannot clarify what those risks are. Or: We want to fund you but an anonymous person vetoed you and we cannot say who or why. Or: Start-up policy projects are too risky. Or: We worry you might be successful and hard to shut down if we decided we don’t like you in future. This makes me worry that the fund managers do not think through the risks and reasons for funding or not funding policy work in as much depth as I would like, and that they maybe do not fund any new/start-up policy projects.
- Acknowledgment that they do apply a higher bar for policy work. Staff at the LTFF have told me that they apply a higher bar for policy work than for other grants.
Of course, this is all circumstantial, and not necessarily a criticism of the LTFF. The fund managers might argue they never get any policy projects worth funding and that all the promising projects I happened to hear of were actually net negative and it was good not to fund them. It is also possible that things have improved in the last year (the notes that make up this post have been sitting in an email chain for a long while now).
Relevance and recommendation
That said I thought it was worth me writing this up publicly as the possibility that the LTFF (and maybe other funders) are systematically not funding any new policy advocacy projects is relevant for:
- Applicants. I don’t want people who have not received funding to be demotivated. To all those folk out there looking to run an EA policy project but have struggled with funding from the LTFF, or elsewhere, I just want to say: Heads up, don’t be put off, the work you do may well be of high value. Stop applying to the LTFF. There are funders more keen on policy such as SFP/SFF or FLI so consider applying there or talking to local funders in your area! Also such founders should feel free to reach out to me if they want an independent view on their project and advice on finding funding.
- Other funders. I don’t want other funders to hold back from funding as they think the LTFF will cover policy stuff. Maybe these funders being aware of exactly how rare it is for the LTFF to fund (new) projects aimed at influencing public policy, will help ensure they are fully informed and can make the best decisions.
- Both funders and policy folk. Maybe there is a discussion to be had here as to why this might be going on. If there are lot of entrepreneurial policy folk applying for funding and not getting it then maybe the entrepreneurial policy folk in EA have a very different view than funders of what kinds of policy projects are valuable. Maybe this is a communication issue and we can bring it into the light. My hope is that a public post could spark some discussion about what kinds of policy projects would bring value to the world, and that this could be of use to both funders and policy entrepreneurs.
Overall, more transparency could help. I would love to hear from funders (LTFF and otherwise) a bit more about what they are looking for. I think this could lead to either push back (why are you looking for that?) or better applications or both. And if the LTFF are not funding policy work, or only funding certain types of policy work, or if they are applying a higher bar to policy projects then they should say this publicly on their website!
Opinions & speculation
I would note that it is only possible to make this post because the LTFF has written public payout reports and given feedback to people who have applied for funding. These are both really good and deserving of praise. There is much the LTFF gets correct, and this critique should be seen in this light. (I also note that my interactions with staff at the LTFF have been extremely positive and I have a lot of respect for them.)
Does this apply to other funders?
I also worry similar problems apply elsewhere in EA but are only noticeable in the LTFF as the LTFF are more transparent and easier to apply to than other funders. Given that senior staff at EA orgs have suggested that there is a general reticence to fund new policy work, it seems plausible that many other (longtermist?) funders are also not funding new policy projects.
A guess as to what might be going on.
I speculate that there is an underfunding of policy projects and that this might be connected to EA funders approach to minimising risks.
EA funders are risk averse and want to avoid funding projects with downside risks. I think this is super important. Most movements that exists fail and fracture and collapse and that EA has succeeded as much as it has is perhaps down to this caution.
However I think:
- Funders and policy actors might have very different views is on how to manage the risks of policy work, meaning they prefer different kinds of projects. Based on the conversations it seems that funders believe that a reasonable plan to manage the risks is to look for very senior (say 20 years experience) that they know very well to start new policy projects. Also they might give veto power to a range of external advisors. On the other hand policy folk think it is more important to be cautious about the kinds of projects started and risk mitigation plans, but they note that senior people often come with political baggage that increases risks whereas junior folk can fly under the political radar. Also they note that giving vetoes has problems as if different people veto different things a funders de-facto bar could be significantly higher than any single person's bar.
- It is not clear to me that funders are considering the risks of not funding projects. Firstly funding can be used by projects to directly decrease risks of that project, especially if it is given with this purpose in mind, such as to hire a good comms person. Secondly, new projects can decrease policy risks, e.g. a project that supports longtermist academics to engage with policy makers believes they have cut not increased the risks. Thirdly, not funding can miss key policy windows, like COVID.
Additionally, there is the risk that a default of not funding action feeds into a culture of EA only ever researching and never acting. If it is the case that many funders are applying a much higher bar to doing/influencing projects than to thinking/research projects (as the LTFF say they do) then this could lead to bad community dynamics. In particular it might be feeding into a culture where doers are ostracised or not supported and thinkers are welcome in with open arms. People have recently been raising this as an issue with EA (or longtermism) culture on this forum see here (see comments too) and here and here.
The above is speculative. It is my current best guess as the crux behind any difference in views between funders and policy folk. That said I would not be surprised if I was wrong on this.
I think both sides have their biases and I would be keen for both sides to talk more on this or I think at a minimum funders should be transparent about their views on funding policy work and minimising risks. If nothing else it could save people wasted time applying for funding they cannot get.
I expect this is worth funders time. I had the impression that as of 2021 a lot of talented policy people who could start new projects were not doing so because of a real or perceived belief that there was a lack of funding. (FTX may have fixed or partly fixed this I don’t know.)
Thank you to the LTFF team and many others for feedback on an early draft and for all the great work they do. Entered as part of the EA red team challenge.
Edit: I have now written up a long list of longtermist policy projects that should be funded, this gives some idea how big the space of opportunities is here: List of donation opportunities (focus: non-US longtermist policy work)
This was a conversation with me, and sadly strikes me as a strawman of the questions asked (I could recount the conversation more in-full, though that feels a bit privacy violating).
Of course on the LTFF I do not take it as a given that any specific approach to improving the long term future will work, and I question actively whether any broad set of approaches has any chance to be competitive with other things that we can do. It's important that the LTFF is capable of coming to the conclusion that policy work is in-general unlikely to be competitive, and just because there are already a bunch of people working in policy, doesn't mean I no longer question whether the broad area might be an uphill battle.
I personally am probably the person on the LTFF most skeptical of policy work, especially of work that aims to do policy-analysis embedded within government institutions, or that looks like it's just a generic "let's try to get smarter people into government" approach. I've seen this fail a good number of times, and many interviews I've done with policy people suggests to me that many people report finding it very hard to think clearly when embedded within government institutions. I also think generic power-seeking behavior where we try to "get our guys into government" is somewhat uncooperative and also has detrimental epistemic effects on the community.
Other people on the LTFF and the rest of the funding ecosystem seem to be more optimistic here, though one thing I've found from discussing my cruxes here for many hundreds of hours with others is that people's models of the policy space drastically differ, and most people are pessimistic about most types of policy work (though then often optimistic about a specific type of policy work).
The questions I actually asked were of the type "how do you expect to overcome these difficult obstacles that seem to generally make policy work ineffective, that seem reported by many people in policy?". I don't think we've had really any policy successes with regards to the Long Term Future, so if a project does not have compelling answers to this kind of question, I am usually uninterested in funding it, though again, others on the LTFF have a pretty different set of cruxes.
I often ask the same types of question with regards to AI Alignment research: "it seems like we haven't really found much traction with doing AI Alignment research, and it seems pretty plausible to me that we don't really know how to make progress in this domain. Why do you think we can make progress here?". I do have a higher prior on at least some AI Alignment research working, and also think the downside risk from marginal bad AI Alignment research is less than from marginal policy advocacy or intervention, so my questions tend to be a bit more optimistically framed, or I can contribute more of the individual model pieces.
Thank you Habryka. Great to get your views.
Apologies if I misrepresented this in some way (if helpful and if you have a recording or notes happy for things I said and you said to be public). I have not had this kind of questioning on any other funding applications and it felt very strange to me. I said in an email to Caleb (EA funds) recently that "it would surprise me a lot if people applying for grants to do say AI safety technical work got an hour of [this type]". So perhaps count me as surprised. If this kind of questioning is just an idoscyratic Habryka tool for getting to better grips with applicants of different types then I am happy with it. Will edit the post.
I guess it depends how specific you are being. Obviously I don’t think it should be taken as given that "specific think tank plan x" would be good, but I do think it is reasonable for a fund to take it as given that at a high level "policy work" would be good. And if the LTFF does not think this then why does the LTFF actively outreach to policy people to get them to apply?
(I recognise there may be a difference here between you Habryka and the rest of LTFF, as you say you are more sceptical than others)
Now it is my turn to claim a strawman. I have never applied to the LTFF with a plan anything close to a "let's try to get smarter people into government" approach. Nor were any of the 5 applications to the LTFF I am aware of anything like this approach.
FWIW, I think this kind of questioning is fairly Habryka-specific and not really standard for our policy applicants; I think in many cases I wouldn’t expect that it would lead to productive discussions (and in fact could be counterproductive, in that it might put off potential allies who we might want to work with later).
I make the calls on who is the primary evaluator for which grants; as Habryka said, I think he is probably most skeptical of policy work among people on the LTFF, and hasn’t been the primary evaluator for almost any (maybe none?) of the policy-related grants we’ve had. In your case, I thought it was unusually likely that a discussion between you and Habryka would be productive and helpful for my evaluation of the grant (though I was interested primarily in different but related questions, not “whether policy work as a whole is competitive with other grants”), because I generally expect people more embedded in the community (and in the case above, you (Sam) in particular, which I really appreciate), to be more open to pretty frank discussions about the effectiveness of particular plans, lines of work, etc.
FWIW I think if this is just how Habryka works then that is totally fine from my point of view. If it helps him make good decisions then great.
(From the unusualness of the questioning approach and the focus on "why policy" I took it to be a sign that the LTFF was very sceptical of policy change as an approach compared to other approaches, but I may have been mistaken in making this assumption based on this evidence.)
I don't think we should take it as a given! I view figuring out questions like this as most of our job, so of course I don't want us to have an institutional commitment to a certain answer in this domain.
In order to believe that something could potentially be furthered by someone, or that it has potential, I don't think I have to take it as a given that work in that general area "would be good".
I also think it's important to notice that the LTFF page only lists "policy research" and "advocacy", and doesn't explicitly list "policy advocacy" or "policy work" more broadly (see Asya's clarification below). I don't think we currently actively solicit a lot of policy work for the LTFF, though maybe other fund managers who are more optimistic about that type of work have done more soliciting.
And separately, the page of course reflects something much closer to the overall funds view (probably with a slant towards Asya, since she is head of the LTFF), and this is generally true for our outreach, and I think it's good and valuable to have people with a diversity of views on the LTFF (and for people who are more skeptical of certain work to talk to the relevant grantees).
Sorry! Seems like this is just me communicating badly. I did not intend to imply (though I can now clearly see how one might read it as such) that your work in-particular falls into this category. I was trying to give some general reasons for why I am skeptical of a lot of policy work (I think only some of these reasons apply to your work). I apologize for the bad communication here.
The page says the LTFF is looking to fund projects on "reducing existential risks through technical research, policy analysis, advocacy, and/or demonstration projects" which a casual reader could take to include policy advocacy. If the aim of the page is to avoid giving the impression that policy advocacy is something the LTFF actively looks to fund then I think it could do a better job.
Maybe this has stopped now. Last posts I saw was Dec 2021 from an EA Funds staff who has now left (here). Posts said things like: "we'd be excited to consider grants related to policy and politics. We fund all kinds of projects" (here). It is plausible to me that things like that were giving people the wrong impressions and the LTFFs willingness to fund certain projects.
Fair enough. That seems like a reasonable approach too and I hope it is going well and you are learning a lot!!
No worries. Sorry too for any ways I have misrepresented our past interactions. Keep being wonderful <3
Thanks for posting this comment, I thought it gave really useful perspective.
This strikes me as an odd statement. If you're talking about the LTF fund, or EA long-termism, it doesn't seem like much policy work has been funded.
If you're talking more broadly, wouldn't policy wins like decreasing the amount of lead being emitted into the atmosphere (which has negative effects on IQ and health generally) be a big policy win for the long term future?
I think this is false, e.g. a reasonable subset of Open Phil's Transformative AI risks grantmaking is on policy.
Huh, why do you think that? CSET was Open Phil's largest grant to date, and I know of at least another $20MM+ in policy projects that have been funded.
Sadly, I think a lot of policy grants are announced less publicly, because publicity is usually harmful for policy projects or positions (which I think is at least some evidence of them being at least somewhat adversarial/non-cooperative, which is one of the reasons why I have a somewhat higher prior against policy projects). Approximately all policy applications to the LTFF end up requesting that we do not publish a public writeup on them, so we often refer them to private funders if we think they are a good idea.
I guess I was just wrong, I hadn't looked into it much!
Bias view incoming .... :
I think LTFF's only (public) historical grant for policy advocacy, to the APPG for Future Generations, has led to better policy in the UK, in particular on risk management. For discussions on this see impact reports here and here, independent reviews here and here, and criticism here.
Additionally I think CLTR has been doing impactful long-term focused policy work in the UK.
Yeah, I think this is definitely a candidate for a great intervention, though I think importantly it wasn't the result of someone entering the policy space with a longtermist mindset.
If someone had a concrete policy they wanted to push for (or a plan for discovering policies) of that magnitude, then I would likely be excited about funding it, though I would still be somewhat worried how likely it would be to differentially accelerate development of dangerous technologies vs. increase humanities ability to navigate rapid technological change (since most risk to the future is anthropogenic, I am generally skeptical of interventions that just speed up technological progress across the board), but my sense is abating lead poisoning looks better than most other things on this dimension.
An offshoot of lead emission in the atmosphere might be the work being done at LEEP (Lead Exposure Elimination Project) https://forum.effectivealtruism.org/posts/ktN29JneoQCYktqih/seven-more-learnings-from-leep
(I work for the LTFF)
Surely something akin to this critique can also be levied at e.g. alignment research.
Oh, sorry, I didn't intend this at all as a critique. I intended this as a way to communicate that I don't think I am that alone in thinking that most policy projects are pretty unlikely to be helpful.
Sorry "critique" was poor choice of words on my part. I just meant "most LT plans will fail, and most LT plans that at least some people you respect like will on an inside view certainly fail" is just the default for trying to reason well on the frontier of LT stuff. But I'm worried that the framing will sound like you meant it narrowly for policy. Also, I'm worried your implied bar for funding policy is higher than what LTFF people (including yourself) actually use.
Hmm, yeah, I think we are both using subpar phrasing here. I think this is true for both policy and AI Alignment, but for example less true for biorisk, where my sense is there is a lot more people agreeing that certain interventions would definitely help (with some disagreement on the magnitude of the help, but much less than for AI Alignment and policy).
I agree about biosecurity, sure. Although, I actually think we're much less conceptually confused about biosecurity policy than we are about AI policy. For example, pushing for a reasonable subset of the Apollo report seems reasonable to me.
Yeah, I think being less conceptually confused is definitely part of it.
Hey, Sam – first, thanks for taking the time to write this post, and running it by us. I’m a big fan of public criticism, and I think people are often extra-wary of criticizing funders publicly, relative to other actors of the space.
Some clarifications on what we have and haven’t funded:
Responding to the rest of the post:
Thank you Abergal, I hope my critique is helpful. I mean it to be constructive.
I don’t think I disagree with anything at all that you wrote here!! So glad we are mostly on the same page.
(In fact you you suggest "we also differ in our views of the upsides of some of this work" and I am not sure that is the case. I am fairly sceptical of much of it, especially more AI focused stuff.)
I still expect the main disagreements are on:
If it is helpful happy to discuss either of these points further.
Also super great to hear that you have made three grants in the last year to projects that interact with policymakers!! Even if this is "very small" compared to other grants it is more than previous years. I look forward to hearing what some of them are if there is a future write-up. :-)
I strongly agree with Sam on the first point regarding downside risks. My view, based on a range of separate but similar interactions with EA funders, is that they tend to overrate the risks of accidental harm  from policy projects, and especially so for more entrepreneurial, early-stage efforts.
To back this up a bit, let's take a closer look at the risk factors Asya cited in the comment above.
To be clear, I do think these risks are all real and worth thinking about! But to my reasonably well-informed understanding of at least three EA grantmakers' processes, most of these projects are not judged by way of a sober risk analysis that clearly articulates specific threat models, assigns probabilities to each, and weighs the resulting estimates of harm against a similarly detailed model of the potential benefits. Instead, the risks are assessed on a holistic and qualitative basis, with the result that many things that seem potentially risky are not invested in even if the upside of them working out could really be quite valuable. Furthermore, the risks of not acting are almost never assessed -- if you aren't trying to get the policymaker's attention tomorrow, who's going to get their ear instead, and how likely might it be that it's someone you'd really prefer they didn't listen to?
While there are always going to be applications that are not worth funding in any grantmaking process, I think when it comes to policy and related work we are too ready to let perfect be the enemy of the good.
Important to note that the observations here are most relevant to policymaking in Western democracies; the considerations in other contexts are very different.
“even if the upside of them working out could really be quite valuable” is the part I disagree with most in your comment. (Again, speaking just for myself), I don’t think any of the projects I remember us rejecting seemed like they had a huge amount of upside; my overall calculus was something like “this doesn’t seem like it has big upside (because the policy asks don’t seem all that good), and also has some downside (because of person/project-specific factors)”. It would be nice if we did quantified risk analysis for all of our grant applications, but ultimately we have limited time, and I think it makes sense to focus attention on cases where it does seem like the upside is unusually high.
On potential risk factors:
Caveating a lot of the above: as I said before, my views on specific grants have been informed heavily by others I’ve consulted, rather than coming purely from some inside view.
Thanks for the response!
That's fair, and I should also be clear that I'm less familiar with LTFF's grantmaking than some others in the EA universe.
Oh, I totally agree that the kind of risk analysis I mentioned is not costless, and for EA Funds in particular it seems like too much to expect. My main point is that in the absence of it, it's not necessarily an optimal strategy to substitute an extreme version of the precautionary principle instead.
Overall, I agree that judging policy/institution-focused projects primarily based on upside makes sense.
Just to be clear, the LTFF generally has a philosophy of funding things if at least one fund member is excited about them, and a high bar for a fund member to veto a grant that someone else is championing, so I don't think "every possible idea getting vetoed" is a very likely failure mode of the LTFF.
The idea that was suggested to me by an EA policy person was not that other fund members veto but that external experts veto.
The hypothetical story is that a fund manager who is not an expert in policy who is struggling to evaluate a policy grant that they worry is high-risk might invite a bunch of external experts to give a view and have a veto and, given the risk, if any of those experts veto they might choose not to fund it.
This then goes wrong if and experts veto for very different reasons (more likely if "there’s a surprising amount of disagreement in what projects [experts] think are good"). In the worse case it could be that almost every policy grant gets vetoed by one expert or another and none get funded. This could significantly raise the bar and it might take a while for a fund manager to notice.
Honestly I have no idea if this actually happens or ever works like this. If this is not how things work then great! Mostly trying to bring out into the open the various tools that people can use to evaluate more risky projects and to flag the potential uses and also downsides for discussion.
Oh, hmm, I think we rarely request feedback from more than one or two experts, so I would be somewhat surprised if this has a large effect. But yeah, definitely in biorisk and policy, I think if the expert we ping tends to have a negative take, we probably don't fund it (and now that you say it, it does feel a bit more like we are asking more experts on policy grants than other grants, so there might be some of the effect that you are describing going on).
If these experts regularly have a large impact on these decisions, that's an argument for transparency about them. This is a factor that could of course be outweighed by other considerations (ability to give frank advice, confidentiality, etc). Perhaps might be worth asking them how they'd feel about being named (with no pressure attached, obviously).
Also, can one volunteer as an expert? I would - and I imagine others (just on this post, perhaps Ian and Sam?) would too.
or funders could, you know, always hire more fund managers who have policy experience!
I don't think hiring for this is easy. You need aligned people with good judgement, significant policy experience, and counterfactual impact that is equal to or worse impact than working for LTFF. Plus, ideally good communication ability and a bit of creativity too.
Sure, but if there are 4 people out there with CS backgrounds who fit the bill there are probably a few without who do too.
The other thing is the idea that “policy” is this general thing seems a little off to me. Someone who knows a thing or two about Congress may not have context or network to evaluate something aimed at some corner of the executive branch, to say nothing of evaluating a policy proposal oriented towards another country.
Hmm I think this makes the problem harder, not easier.
Was this always true? I (perhaps arrogantly) thought that there was a push for greater grantmaker discretion after my comment here (significantly before I joined LTFF), though it was unclear if my comment had any causal influence
I think we always had a culture of it, but I do think your comment had a causal influence on us embedding that more into the decision-making process.
I hope you don’t mind me returning to this post. I wanted to follow up on a few things but it has been a super busy month and I have not really had the time.
Essentially I think we did not manage to engage with each others cruxes at all. I said I expect the crux is "connected to EA funders approach to minimising risks." You did not discuss this at all in your reply (except to agree that there are downsides to policy work). But then you suggest you maybe a crux is "our views of the upsides of some of this work". And then I replied to you but did not discuss that point at all!!
So overall it felt a bit like we just talked past each other and didn’t really engage fully with the others points of view. So I felt I should just come back and respond to you about the upsides rather than leave this as was.
Some views I have (at least from a UK perspective, cannot comment on US):
Anyway I hope that is useful and maybe gets a bit more at some of the cruxes. Would be keen to hear views on if you think any of this is correct, even if you don’t have time to respond in depth to any of these points.
Thank you so much for engaging and listening to my feedback – good luck with future grantmaking!!!
Thank you for sharing, Abergal!
I was just wondering if you could share more concrete examples of "taking up the space" risks. We're facing some choices around this in Australia at the moment and I want to make sure we've considered all downsides of uniting under a shared vision. Are the risks of "taking up the space" mainly:
I'd be keen to know if there are others we may not have considered.
All four current fund managers at LTFF have degrees in computer science, and none have experience in policy. Similarly, neither of OpenPhil's two staff members on AI Governance have experience working in government or policy organizations. These grantmakers do incredible work, but this seems like a real blind spot. If there are ways that policy can improve the long-term future, I would expect that grantmakers with policy expertise would be best positioned find them.
EDIT: See below for the new LTFF grantmaker with exactly this kind of experience :)
This is not up to date, sorry! I think it should be updated in the next few days.
Ah, no worries. Are there any new grantmakers with policy backgrounds?
Rebecca Kagan is currently working as a fund manager for us (sorry for the not-up-to-date webpage).
That's really cool! Seems like exactly the kind of person you'd want for policy grantmaking, with previous experience in federal agencies, think tanks, and campaigns. Thanks for sharing.
Yep great news!
Is Rebecca still a fund manager, or is the LTFF page out of sync?
(I work for the LTFF)
I think there's three implicit assumptions here that I might disagree with:
Re: 2, I don't think we have strong reasons to think this is true. Anecdotally, I have more of a forecasting background than other LTFF fund managers, and I think I am not more optimistic about forecasting grants than the rest of LTFF.
There's a version of your comment that I do agree with, which is that all else equal it's great for fund managers to have a diversity of contacts and experiences, especially in areas of projects that we good applicants in. To that end, I'm a bit worried that none of current fund managers have done much object-level work in AI alignment, or to a lesser extent technical biosecurity*.
*In recent months I ended up being the primary investigator for most bio grants. I don't feel experienced enough to be very happy about my judgements, but I think I know enough high-level context and have enough contacts that it's mostly been going okay. This problem is ameliorated somewhat by having only so many biosecurity applications in the first place. Please consider pinging me, Asya, or Caleb if you have feedback on any of the grants I investigated.
I basically agree with this. On 1, undergrad degrees aren’t a great proxy but particularly the people listed on the LTFF site are all career engineers. On 2, your description sounds like the correct general case, but in a case where non-policy people are questioning the effectiveness of any policy work on the grounds that policy is ineffective, I would expect people who’d worked on it to usually have a brighter view given that they’ve chosen to work on it. 3 is of course up for debate and the main question.)
As someone with a fair amount of context on longtermist AI policy-related grantmaking that is and isn't happening, I'll just pop in here briefly to say that I broadly disagree with the original post and broadly agree with [https://forum.effectivealtruism.org/posts/Xfon9oxyMFv47kFnc/some-concerns-about-policy-work-funding-and-the-long-term?commentId=TEHjaMd9srQtuc2W9](abergal's reply).
Hey Luke. Great to hear from you. Also thank you for your push back on an earlier draft where I was getting a lot of stuff wrong and leaping to silly conclusions, was super helpful. FWIW I don’t know how much any of this applies to OpenPhil.
Just to pin down on what it is you agree / disagree with:
For what it is worth I also broadly agree with abergal's reply. The tl;dr of both the original post / abergal's comment is basically the same: hey it [looks like from the outside/ is the case that] the LTFF is applying a a much higher bar to direct policy work than other work.
I guess the original post also said: Hey we might be leaving a bunch of value on the table with this approach, and here are a few reasons why, and that is a thing to think about. Abergal's reply did not directly address this although I guess it implied that Abergal is happy with the current status quo and/or doesn't see value in exploring this topic. So maybe that is the thing you are saying you agree with.
Note to anyone still following this: I have now written up a long list of longtermist policy projects that should be funded, this gives some idea how big the space of opportunities is here: List of donation opportunities (focus: non-US longtermist policy work)