I live for a high disagree-to-upvote ratio
Thank you so much! Your criticism has helped me identify a few mistakes, and I think can get us closer to clarity. The main difference between our models is around who counts as a 'beneficiary', or what it means to 'recruit' someone.
The main thing I want to focus on is that you're predicting a cost per beneficiary that would be nearly 50% recruitment. I don't think that passes the smell test. The main difference is you're only counting the staff time for active participants, but even with modest dropout, we'd expect the vast majority of staff time to go to users who only complete one or two calls. But you're right to point out that we should factor in dropout between installation and the first guidance call, and when I factor this in, unguided has 11% of the cost of guided at scale.
The rest of this comment is just my working out.
One of the mistakes I made was having different definitions for recruitment for each condition in my cost model. If the number in that model says there are 100,000 beneficiaries, in the unguided model this means we got 100,000 installs, but in the guided model it means we got 100,000 participants who had 50–180 minutes of staff time allocated to them. Obviously there are different costs to recruit these kinds of participants.
(Two other mistakes I found: I shouldn't have multiplied the unguided recruitment cost by 2 to account for engagement differences, and I forgot to discount the office space costs for the nonexistent guides in the unguided model)
To rectify this, let's count a 'beneficiary' as someone who is in the targeted subgroup and completes pre-treatment. This is in line with most of the literature, which counts 'dropout' regardless of whether users complete any of the material, so long as they've done their induction. We don't want to filter this down to 'active' users, since users who drop out will still incur costs.
We have some facts from Kaya Guides:
I think this fairly accounts for everything you raised. I think you're right to point out that my model should've accounted for the cost of a treatment starter (~8x higher). But I don't think it's right to only account for active users, since Kaya Guides spend 15 minutes of staff time on 12% of all installers, even if they drop out later. And as their ad targeting gets better, we'd only expect this number to increase, paradoxically widening the cost gap!
Plugging it all in, unguided has 8% (17–23%) of the cost of guided at scale.
Earlier, I also sense-checked with Kaya Guides' direct cost-per-beneficiary, which they estimate to be $3.93. If the unguided cost per beneficiary is $0.41 (as in the updated model), then the limiting proportion increases a bit to 11%.
The fact that it took over a month to find some pretty obvious flaws in my model is a concern, and my model is clearly somewhat sensitive to the parameters. However, even if I'm really pessimistic about the parameters, I can't get it above 20% of the cost of guided, which would still make it more cost-effective.
The bigger doubt I've had since writing this report is learning from Kaya Guides that they actually do have an unguided condition—anyone who scores 0–9 on the PHQ-9 (no/mild depression), or anyone who scores above but explicitly doesn't want a guide gets the ordinary programme, just without the calls. This has an astonishing 0% completion rate. I think the different subgroup, programme design, and lack of focus are mostly contributing to this, but it indicates that it's gonna be hard to keep users engaged. I'll chat with them some more and see if I can learn anything else.
My union is pretty conservative w/r/t social justice, because it's the one that covers tech & science (our members tend to hold left-wing opinions, but don't like stirring the pot). I don't know how we'd feel about animal welfare, but not many of us work directly in those industries.
To get closer to your point, live animal export is a big issue in Australia, and our dedicated Meat Industry Employees Union have called for a ban on it. So I think the kind of campaign you're talking about would fit right in here. Their animal welfare policy is so important to them that it's on the front page of their website. Equally, they've worked with the Greens and the Animal Justice Party (both legislatively represented) in the past, and the unions here have close ties to the Labor party (1 of 2 major parties), so political change might be uniquely achievable here—although I doubt the situation is much different in most EU countries.
Thank you—I am a big believer in the power of collective action & have organised successful union drives & pay disputes in the past. I don't have a lot to add to your breakdown; I think this is a very promising area for EA to consider for almost every cause area (ex. would love to see a similar breakdown for current/future efforts in frontier AI labs).
Just strategically, I think the most promising insider activism campaign would be to partner with an existing union in a country with strong union protections; this way, you can leverage those protections to prevent retaliation against employee activists, as they can credibly claim they were organising for the union. I think, frankly, this rules out the U.S. as a starting point—you would want to build groundswell in places where the host companies can't cut out off at the knees (the recent dismissals at Google are a strong reminder that if employees protest something the company has a stake in, they'll be fired at-will with no consequences).
Furthermore, unions have a lot of existing connections & skills in developing these campaigns, and, as you've noted, regularly participate in employee activism directly or otherwise have presences in other social movements. This comes with the trade-off of potentially alienating some employees (unions are almost exclusively left-wing and have established reputations), but I don't think there are many people (outside of the U.S.) who would be put off by a union and would've otherwise joined an employee activist drive.
FWIW on timelines:
Without reading too much into it, there's a similar amount of negativity about the state of EA as there is a lack of confidence in its future. That suggests to me that there's a lot of people who think EA should be reformed to survive (rather than 'it'll dwindle and that's fine' or 'I'm unhappy with it but it'll be okay')?
Global Burden of Disease (GBD) is okay, it depends a lot on what disease & metric you're looking at, and how aware you are of the caveats around it. Some of these:
Overall, I think the GBD is very robust and an extremely useful tool, especially for (a) making direct comparisons between countries or diseases and (b) where no direct, trustworthy, country-specific data is available. But you should be able to improve on its accuracy if you have an inside view on a particular situation. I don't think it's subject to the incentives you mention above in quite the same way.