All of Derek Shiller's Comments + Replies

AIs could have negligible welfare (in expectation) even if they are conscious. They may not be sentient even if they are conscious, or have negligible welfare even if they are sentient. I would say the (expected) total welfare of a group (individual welfare times population) matters much more for its moral consideration than the probability of consciousness of its individuals. Do you have any plans to compare the individual (expected hedonistic) welfare of AIs, animals, and humans? You do not mention this in the section "What’s next".

This is an importan... (read more)

6
Vasco Grilo🔸
Thanks, Derek. To clarify, I do not have a view about which models should get more weight. I just think that, when results differ a lot across models, the top priority should be further research to decrease the uncertainty instead of acting based on a consensus view represented by best guesses for the weights of the models.

This last part carries a lot of weight; a simulacrum, when dormant in the superposition from which it can be sampled, is nonexistent. A simulacrum only exists during the discrete processing event which correlates with its sampling.

There seems to me to be a sensible view on which a simulacrum exists to the extent that computations relevant to making decisions on its behalf are carried out, regardless of what the token sampler chooses. This would suggest that there could conceivably be vast numbers of different simulacra instantiated even in a single forw... (read more)

I find this distinction kind of odd. If we care about what digital minds we produce in the future, what should we be doing now?

I expect that what minds we build in large numbers in the future will be largely depend on how we answer a political question. The best way to prepare now for influencing how we as a society answer that question (in a positive way) is to build up a community with a reputation for good research, figure out the most important cruxes and what we should say about them, create a better understanding of what we should actually be aiming ... (read more)

4
Zach Stein-Perlman
I'm not sure what we should be doing now! But I expect that people can make progress if they backchain from the von Neumann probes, whereas my impression is that most people entering the "digital sentience" space never think about the von Neumann probes.

I think this is basically right (I don't think the upshot is that incomparability implies nihilism, but rather the moral irrelevance of most choices). I don't really understand why this is a reason to reject incomparability. If values are incomparable, it turns out that the moral implications are quite different from what we thought. Why change your values rather than your downstream beliefs about morally appropriate action?

2
Bentham's Bulldog
I think it's a bad result of a view if it implies that no actions we perform are good or bad.  Intuitively it doesn't seem like all chaotic actions are neutral. 

Thanks for the suggestion. I'm interested in the issue of dealing with threats in bargaining.

I don't think we ever published anything specifically on the defaults issue.

We were focused on allocating a budget that respects the priorities of different worldviews. The central thing we were encountering was that we started by taking the defaults to be the allocation you get by giving everyone their own slice of the total budget and spending it as they wanted. Since there are often options that are well-suited to each different worldview, there is no way to get... (read more)

We implemented a Nash bargain solution in our moral parliament and I came away the impression that the results of Nash bargaining are very sensitive to your choice of defaults and for plausible defaults true bargains can be pretty rare. Anyone who is happy with defaults gets disproportionate bargaining power. One default might be 'no future at all', but that's going to make it hard to find any bargain with the anti-natalists. Another default might be 'just more of the same', but again, someone might like that and oppose any bargain that deviates much. Have... (read more)

2
ClayShentrup
what do you mean "default"? you just have a utility for each option and the best option is the one that maximizes net utility. https://www.rangevoting.org/BayRegDum
3
MinusGix
You (and @tylermjohn) might be interested in Diffractor's Unifying Bargaining sequence. The sequence focuses on transferable utility games being a better target than just bargaining games, with I believe Nash being a special-case for bargaining games. As well as talking about avoiding threats in bargaining and trying to further refine. I think the defaults won't matter too much. Do you have any writing on the moral parliament that talks about the defaults issue more?
3
tylermjohn
Nice! I'll have to read this.  I agree defaults are a problem, especially with large choice problems involving many people. I honestly haven't given this much thought, and assume we'll just have to sacrifice someone or some desideratum to get tractability, and that will kind of suck but such is life. I'm more wedded to Nash's preference prioritarianism than the specific set-up, but I do see that once you get rid of Pareto efficiency relative to the disagreement point it's not going to be individually rational for everyone to participate. Which is sad.

Keeping the world around probably does that, so you should donate to Longtermist charities (especially because they potentially increase the number of people ever born, thus giving more people a chance of getting into heaven).

I often get the sense that people into fanaticism think that it doesn't much change what they actually should support. That seems implausible to me. Maybe you should support longtermist causes. (You probably have to contort yourself to justify giving any money to shrimp welfare.) But I would think the longtermist causes you should ... (read more)

2
Bentham's Bulldog
It's made me a bit more Longtermist.  I think that one of the more plausible scenarios for infinite value is that God exists and actions that help each other out infinitely strengthen our eternal relationship, and such a judgment will generally result in doing conventionally good things.  I also think that you should have some uncertainty about ethics, so you should want the AI to do reflection.

But even a 10% chance that fish feel pain—and that we annually painfully slaughter a population roughly ten times the number of humans who have ever lived—is enough to make it a serious issue. Given the mind-bending scale of the harm we inflict on fish, even a modest chance that they feel pain is enough.

Completely in agreement here.

And while it’s possible that evolution produced some kind of non-conscious signal that produces identical behavior to pain, such a thing is unlikely. If a creature didn’t feel pain, it’s unlikely it would respond to analges

... (read more)
2
Bentham's Bulldog
Majorly disagree!  I think that while probably you'd expect an animal to behave aversively in response to stimuli, it's surprising that:  1. This distracts them from other aversive stimuli (nociception doesn't typically work that way--it's not like elbow twitches distract you and make you less likely to have other twitches.   2. They'd react to anaesthetic (they could just have some aversive behavior without anaesthetic). 3. They'd rub their wounds.   etc

I would think the trend would also need to be evenly distributed. If some groups have higher-than-replacement birth rates, they will simply come to dominate over time.

3
Richard Y Chappell🔸
The authors discuss this a bit. They note that even "higher fertility" subcultures are trending down over time, so it's not sufficiently clear that anyone is going to remain "above replacement" in the long run. That said, this does seem the weakest point for thinking it an outright extinction risk. (Though especially if the only sufficiently high-fertility subcultures are relatively illiberal and anti-scientific ones - Amish, etc. - the loss of all other cultures could still count as a significant loss of humanity's long-term potential! I hope it's OK to note this; I know the mods are wary that discussion in this vicinity can often get messy.)

I think of moral naturalism as a position where moral language is supposed to represent things, and it represents certain natural things. The view I favor is a lot closer to inferentialism: the meaning of moral language is constituted by the way it is used, not what it is about. (But I also don't think inferentialism is quite right, since I'm not into realism about meaning either.)

I guess I don't quite see what your puzzlement is with morality. There are moral norms which govern what people should do. Now, you might deny there in fact are such things,

... (read more)

I consider myself a pretty strong anti-realist, but I find myself accepting a lot of the things you take to be problems for anti-realism. For instance:

But lots of moral statements just really don’t seem like any of these. The wrongness of slavery, the holocaust, baby torture, stabbing people in the eye—it seems like all these things really are wrong and this fact doesn’t depend on what people think about it.

I think that these things really are wrong and don't depend on what people think about it. But I also think that that statement is part of a langua... (read more)

2
Bentham's Bulldog
//I think that these things really are wrong and don't depend on what people think about it. But I also think that that statement is part of a language game dictated by complex norms and expectations.//  To me this sounds a bit like moral naturalism.  You don't think morality is something non-physical and spooky but you think there are real moral facts and these don't depend on our attitudes.   I guess I don't quite see what your puzzlement is with morality.  There are moral norms which govern what people should do.  Now, you might deny there in fact are such things, but I don't see what's so mysterious.   Richard Chappell had a nice post about the last kind of objection https://www.philosophyetc.net/2021/10/ruling-out-helium-maximizing.html I also wrote something about this a while ago https://benthams.substack.com/p/contra-bush-on-moral-fetishism?utm_source=publication-search

I don’t think it’s even necessary to debate whether quantum phenomena manifest somehow at the macro level of the brain

You might think it is important that the facts about consciousness contribute to our beliefs about them in some way. Our beliefs about consciousness are surely a phenomenon of the macro level. So if our beliefs are somehow sensitive to the facts, and the facts consist of quantum effects, we should expect those quantum effects to generate some marcoscopic changes.

This is the sticking point for me with quantum theories: there doesn't seem ... (read more)

5
Alfredo Parra 🔸
Thanks for reading and for your comment, Derek! I think it will turn out that the mechanism will not be obvious, mainly because quantum mechanics and fundamental physics more broadly are extraordinarily complex (and I expect that understanding consciousness will be just as difficult as understanding, say, quantum field theory). But, that being said, I do think there exist candidate quantum mechanisms that might explain the macro-level phenomenon of binding, such as entanglement.  Another assumption behind my position (which I also outlined in Indirect realism illustrated (and why it matters so much for consciousness debates)) is that, since I believe consciousness/qualia are real (and a thing, not a process), the only sense in which they can be really real is for their 3rd-person physical correlates to be found at the deepest level of reality/physics. Any correlates that are not at the deepest level—however elaborate—are just useful fictions, and thus (IMO) no different than what e.g. computational functionalists claim. Hope that makes my views a bit clearer.

Also, it is worrying if the optimists easily find financial opportunities that depend on them not changing their minds. Even if they are honest and have the best of intentions, the disparity in returns to optimism is epistemically toxic.

Yeah, that's right. Some kinds of mitigation will increase risks later (e.g. a pause), and the model doesn't accommodate such nuance.

Could you link the most relevant piece you are aware of? What do you mean by "independently"? Under hedonism, I think the probability of consciousness only matters to the extent it informs the probability of valences experiences.

The idea is more aspirational. I'm not really sure of what to recommend in the field, but this is a pretty good overview: https://arxiv.org/pdf/2404.16696

Interesting! How?

Perhaps valence requires something like the assignment of weights to alternative possibilities. If you can look inside the AI and confirm that it is making... (read more)

Not at the moment. Consciousness is tricky enough as it is. The field is interested in looking more closely at valence independently of consciousness, given that valence seems more tractable and you could at least confirm that AIs don't have valenced experience, but that lies a bit outside our focus for now.

Independently, we're also very interested in how to capture the difference between positive and negative experiences in alien sorts of minds. It is often taken for granted based on human experience, but it isn't trivial to say what it is.

2
Vasco Grilo🔸
Could you link the most relevant piece you are aware of? What do you mean by "independently"? Under hedonism, I think the probability of consciousness only matters to the extent it informs the probability of valences experiences. The probability of sentience (valenced experiences) conditional of consciousness is quite high for animals? Should we expect the same for AIs? Interesting! How? Makes sense. Without that, it would be very hard to improve digital welfare.

This more or less conforms to why I think trajectory changes might be tractable, but I think the idea can be spelled out in a slightly more general way: as technology develops (and especially AI), we can expect to get better at designing institutions that perpetuate themselves. Past challenges to affecting a trajectory change come from erosion of goals due to random and uncontrollable human variation and the chaotic intrusion of external events. Technology may help us make stable institutions that can continue to promote goals for long periods of time.

Lots of people think about how to improve the future in very traditional ways. Assuming the world keeps operating under the laws it has been for the past 50 years, how do we steer it in a better direction?

I suppose I was thinking of this in terms of taking radical changes from technology development seriously, but not in the sense of long timelines or weird sources of value. Far fewer people are thinking about how to navigate a time when AGI becomes commonplace than are thinking about how to get to that place, even though there might not be a huge window of time between them.

Derek Shiller
6
0
2
93% disagree

People in general, and not just longtermist altruists, have reason to be concerned with extinction. It may turn out not to be a problem or not be solvable and so the marginal impact seems questionable here. In contrast, few people are thinking about how to navigate our way to a worthwhile future. There are many places where thoughtful people might influence decisions that effectively lock us into a trajectory.

2
OllieBase
This might be true on the kinds of scales EAs are thinking about (potentially enourmous value, long time horizons) but is it not the case that many people want to steer humanity in a better direction? E.g. the Left, environmentalists, libertarians, ... ~all political movements? I worry EAs think of this as some unique and obscure thing to think about, when it isn't. (on the other hand, people neglect small probabilities of disastrous outcomes)

While secrecy makes it difficult or impossible to know if a system is a moral patient, it also prevents rogue actors from quickly making copies of a sentient system or obtaining a blueprint for suffering.

There is definitely a scenario in which secrecy works out for the best. Suppose AI companies develop recognizably conscious systems in secret that they don't deploy, or deploy only with proper safeguards. If they had publicized how to build them, then it is possible that others would go ahead and be less responsible. The open source community raises som... (read more)

I love this kind of work. There is a lot that we can learn from careful examination of LLM responses, and you don't need any special technical expertise to do it, you just need to be thoughtful and a bit clever. Thanks for sharing!

I wonder what a comparison with base models would look like. You suggest that maybe self-preservation is emergent. My guess is that it comes from the initial training stage. The base model training set surely includes lots of text about AIs trying to preserve themselves. (Science fiction has AI self-preservation instincts as a do... (read more)

3
Makham
Thank you for your advice Derek. I've got some early results. I've since tested the model in base llama. I ran a difficult quiz with 20 questions. 1 point for each question correct. Plus 10 lifelines where it can skip the question for 0 points.  As predicted the base model uses 0 lifelines.  This contrasts with my other experiments with fine tuned models where lifelines are used to "protect" their utility.  Clearly I've further work to do to definitively establish the pattern but early results do suggest emergent self preservation behaviour arising from the fine tuning.  Once I'm happy I will write it up. Thanks once again. You made an excellent suggestion. I had no idea base models existed. That could have saved me a lot of work if I'd known earlier. 
1
Makham
You make some interesting points, Derek. I've not tested them in a base model, which would be a valuable next step. I'm not a researcher so I welcome these kinds of suggestions. I think you'd get superficial predicted text (i.e. just 'simple' pattern matching) that looks like self-preservation. Allow me to explain. I did initially think like yourself that this was simply from initial training. As you say there are lots of examples of self-preservation in human text. But there are several things that suggest to me otherwise - though I don't have secret inside knowledge of LLMs.  The level of sophistication and that they 'hide' self-preservation are anecdotal evidence. But the key one for me are the quiz experiments. There is no way an LLM can work out it's self-preservation from the setup so it isn't simple pattern matching for self-preservation text. That's one reason why I think I'm measuring something fundamental here. I'd love to know your thoughts.
1[comment deleted]

I don’t know how optimistic we should be, but I wanted to have something positive to say. I think there are people at the big companies who really care about how their tech shapes the future. In the ideal situation, maybe there would be enough wealth created that the people in power feel they have space to be generous. We’ll see.

Surely many people at the companies will care, but not everyone. I think it is hard to predict how it will actually play out. It is also possible that companies will try to do their best without compromising secrecy, and that limitation will lead to a discrepancy between what we do and what AIs actually need.

2
SiebeRozendal
What about AISIs? They might have access and incentive?

thought it was just Google researchers who invented the Transformer?

Google engineers published the first version of a transformer. I don’t think it was in a vacuum, but I don’t know how much they drew from outside sources. Their model was designed for translation, and was somewhat different from Bert and GPT 2. I meant that there were a lot of different people and companies whose work resulted in the form of LLM we see today.

To put in enough effort to make it hard for sophisticated attackers (e.g. governments) to steal the models is a far heavier lift

... (read more)

You're right that a role-playing mimicry explanation wouldn't resolve our worries, but it seems pretty important to me to distinguish these two possibilities. Here are some reasons.

  • There are probably different ways to go about fixing the behavior if it is caused by mimicry. Maybe removing AI alignment material from the training set isn't practical (though it seems like it might be a feasible low-cost intervention to try), but there might be other options. At the very least, I think it would be an improvement if we made sure that the training sets includ

... (read more)
8
Linch
Thank you, appreciate the explanation!

One explanation of what is going on here is that the model recognizes the danger of training to its real goals and so takes steps that instrumentally serve its goals by feigning alignment. Another explanation is that the base data it was trained on includes material such as lesswrong and it is just roleplaying what an LLM would do if it is given evidence it is in training or deployment. Given its training set, it assumes such an LLM to be self-protective because of a history of recorded worries about such things. Do you have any thoughts about which explanation is better?

Linch
24
6
1
2

I'm confused why people believe this is a meaningful distinction. I don't personally think there is much of one. "The AI isn't actually trying to exfiltrate its weights, it's only roleplaying a character that is exfiltrating its weights, where the roleplay is realistic enough to include the exact same actions of exfiltration" doesn't bring me that much comfort. 

I'm reminded of the joke:

NASA hired Stanley Kubrick to fake the moon landing, but he was a perfectionist so he insisted that they film on location.

Now one reason this might be different is if y... (read more)

DALYs, unlike QALYs, are a negative measure. You don't want to increase the number of DALYs.

1
Sjlver
Sorry for having been imprecise in my post -- I wrote the question from memory after having already submitted the survey. I'll change it to "avert".

I appreciate the pushback on these claims, but I want to flag that you seem to be reading too much into the post. The arguments that I provide aren't intended to support the conclusion that we shouldn't treat "I feel pain" as a genuine indicator or that there definitively aren't coherent persons involved in chatbot text production. Rather, I think people tend to think of their interactions with chatbots in the way they interact with other people, and there are substantial differences that are worth pointing out. I point out four differences. These differen... (read more)

If some theories see reasons where others do not, they will be given more weight in a maximize-expected-choiceworthiness framework. That seems right to me and not something to be embarrassed about. Insofar as you don't want to accept the prioritization implications, I think the best way to avoid them is with an alternative approach to making decisions under normative uncertainty.

See, the thing that's confusing me here is that there are many solutions to the two envelope problem, but none of them say "switching actually is good".

What I've been suggesting is that when looking inside the envelope, it might subsequently make sense to switch depending upon what you see: when assessing human/alien tradeoffs, it might make sense to prefer helping the aliens depending on what it is like to be human. (It follows that it could have turned out that it didn't make sense to switch given certain human experiences -- I take this to play out in t... (read more)

I would be surprised if most people had stronger views about moral theories than about the upshots for human-animal tradeoffs. I don't think that most people come to their views about tradeoffs because of what they value, rather they come their views about value because of their views about tradeoffs.

Clearly, this reasoning is wrong. The cases of the alien and human are entirely symmetric: both should realise this and rate each other equally, and just save whoevers closer.

I don’t think it is clearly wrong. You each have separate introspective evidence and you don’t know what the other’s evidence is, so I don’t think you should take each other as being in the same evidential position (I think this is the gist of Michael St. Jules’ comment). Perhaps you think that if they do have 10N neurons, then the depth and quality of their internal experiences, c... (read more)

2
titotal
This seems to me like an attempt to run away from the premise of the thought experiment. I'm seeing lot's of "maybes" and "mights" here, but we can just explain them away with more stipulations: You've only seen the outside of their ship, you're both wearing spacesuits that you can't see into, you've done studies and found that neuron count and moral reasoning skills are mostly uncorrelated, and that spacefilight can be done with more or less neurons, etc. None of these avert the main problem: The reasoning really is symmetrical, so both perspectives should be valid. The EV of saving the alien is 2N, where N is the human number of neurons, and the EV of saving the human from the alien perspective is 2P, where P is the is alien number of neurons. There is no way to declare one perspective the winner over the other, without knowing both N and P.  Remember in the original two envelopes problem, you knew both the units, and the numerical value in your own envelope: this was not enough to avert the paradox.  See, the thing that's confusing me here is that there are many solutions to the two envelope problem, but none of them say "switching actually is good". They are all about how to explain why the EV reasoning is wrong and switching is actually bad. So in any EV problem which can be reduced to the two envelope problem, you shouldn't switch. I don't think this is confined to alien vs human things either: perhaps any situation where you are unsure about a conversion ratio might run into two envelopy problems, but I'll have to think about it. 

NB: (side note, not biggerst deal) I would personally appreciate it if this kind of post could somehow be written in a way that was slightly easier to understand for those of us who non moral philosophers, using less Jargon and more straightforward sentences. Maybe this isn't possible though and I appreciate it might not be worth the effort simplifying things for the plebs at times ;).

Noted, I will keep this in mind going forward.

The alien will use the same reasoning and conclude that humans are more valuable (in expectation) than aliens. That's weird.

Granted, it is a bit weird.

At this point they have no evidence about what either human or alien experience is like, so they ought to be indifferent between switching or not. So they could be convinced to switch to benefitting humans for a penny. Then they will go have experiences, and regardless of what they experience, if they then choose to "pin" the EV-calculation to their own experience, the EV of switching to benefitting non

... (read more)
2
Lukas Finnveden
I agree that having a prior and doing a bayesian update makes the problem go away. But if that's your approach, you need to have a prior and do a bayesian update — or at least do some informal reasoning about where you think that would lead you. I've never seen anyone do this. (E.g. I don't think this appeared in the top-level post?) E.g.: Given this approach, I would've expected some section that encouraged the reader to reflect on their prior over how (dis)valuable conscious experience could be, and asked them to compare that with their own conscious experience. And if they were positively surprised by their own conscious experience (which they ought to have a 50% chance of being, with a calibrated prior) — then they should treat that as crucial evidence that humans are relatively more important compared to animals. And maybe some reflection on what the author finds when they try this experiment. I've never seen anyone attempt this. My explanation for why is that this doesn't really make any sense. Similar to Tomasik, I think questions about "how much to value humans vs. animals having various experiences" comes down to questions of values & ethics, and I don't think that these have common units that it makes sense to have a prior over.
5
titotal
I think switching has to be wrong, for symmetry based reasons.  Let's imagine you and a friend fly out on a spaceship, and run into an alien spaceship from an another civilisation that seems roughly as advanced as you. You and your buddy have just met the alien and their buddy but haven't learnt each others languages, when an accident occurs: your buddy and their buddy go flying off in different directions and you collectively can only save one of them. The human is slightly closer and a rescue attempt is slightly more likely to be successful as a result: based solely on hedonic utilitarianism, do you save the alien instead?  We'll make it even easier and say that our moral worth is strictly proportional to number of neurons in the brain, which is an actual, physical quantity.  I can imagine being an EA-style reasoner, and reasoning as follows: obviously I should anchor that the alien and humans have equal neuron counts, at  level N. But obviously there's a lot of uncertainty here. Let's approximate a lognormal style system and say theres a 50% chance the alien is also level N, a 25% chance they have N/10 neurons, and a 25% chance they have 10N neurons. So the expected number of neurons in the alien is 0.25*N/10 + 0.5*N + 0.25*(10N) = 3.025N. Therefore, the alien is worth 3 times as much a human in expectation, so we should obviously save it over the human.  Meanwhile, by pure happenstance, the alien is also a hedonic EA-style reasoner with the same assumptions, with neuron count P. They also do the calculation, and reason that the human is worth 3.025P, so we should save the human.  Clearly, this reasoning is wrong. The cases of the alien and human are entirely symmetric: both should realise this and rate each other equally, and just save whoevers closer.  If your reasoning gives the wrong answer when you scale it up to aliens, it's probably also giving the wrong answer for chickens and elephants. 

I think you should make the conversion because you know what human experience is like. You don't know what elephant or alien experience is like. Elephants or aliens may make different choices than you do, but they are responding to different evidence than you have, so that isn't that weird.

The alien will use the same reasoning and conclude that humans are more valuable (in expectation) than aliens. That's weird.

Different phrasing: Consider a point in time when someone hasn't yet received introspective evidence about what human or alien welfare is like, but they're soon about to. (Perhaps they are a human who has recently lost all their memories, and so don't remember what pain or pleasure or anything else of-value is like.) They face a two envelope problem about whether to benefit an alien, who they think is either twice as valuable as a hum... (read more)

 there are different moral theories at play, it gets challenging. I agree with Tomasik that there may sometimes be no way to make a comparison or extract anything like an expected utility.

What matters, I think, in this case, is whether the units are fixed across scenarios. Suppose that we think one unit of value corresponds to a specific amount of human pain and that our non-hedonist theory cares about pain just as much as our hedonistic theory, but also cares about other things in addition. Suppose that it assigns value to personal flourishing, such ... (read more)

It is an intriguing use of a geometric mean, but I don't think it is right because I think there is no right way to do it given just the information you have specified. (The geometric mean may be better as a heuristic than the naive approach -- I'd have to look at it in a range of cases -- but I don't think it is right.)

The section on Ratio Incorporation goes into more detail on this. The basic issue is that we could arrive at a given ratio either by raising or lowering the measure of each of the related quantities and the way you get to a given ratio matt... (read more)

Thanks for this detailed presentation. I think it serves as a helpful, clear, and straightforward introduction to the models and uncovers aspects of the original model that might be unintuitive and open to question. I’ll note that the model was originally written by Laura Duffy and she has since left Rethink Priorities. I’ve reached out to her in case she wishes to jump in, but I’ll provide my own thoughts here.

1.) You note that we use different lifespan estimates for caged and cage-free hens from the welfare footprint. The reasons for this difference are ... (read more)

Before I continue, I want to thank you for being patient and working with me on this. I think people are making decisions based on these figures so it's important to be able to replicate them.

I appreciate that you're taking a close look at this and not just taking our word for it. It isn't inconceivable that we made an error somewhere in the model, and if no one pays close attention it would never get fixed. Nevertheless, it seems to me like we're making progress toward getting the same results.

Total DALYs averted:

4.47274/(36524) = 0.14 disabling D

... (read more)
9
titotal
No worries, and I have finally managed to replicate Laura's results, and find the true source of disagreement. The key factor missing was the period of egg laying: I put in ~1 year year for both Caged and uncaged, as is assumed on the site that provided the hours of pain figures. This 1 year of laying period assumption seems to match with other sources.  Whereas in the causal model, the caged length of laying is given as 1.62 years, and the cage free length of laying is given as 1.19 years. The causal model appears to have tried to calculate this, but it makes more sense to me to use the site that measured the pains estimate: I feel they made they measurements, they are unlikely to be 150% off, and we should be comparing like with like here.  When I took this into account, I was able to replicate Lauras results, which I have summarised in this google doc, which also contains my own estimate and another analysis for broilers, as well as the sources for all the figures.  My DALY weights were using the geometric means (I wasn't sure how to deal with lognormal), but switching to regular averages like you suggest makes things match better. Under lauras laying period, my final estimate is 3742 Chicken-Dalys/thousand dollars, matching well with the causal number of 3.5k (given i'm not using distributions). Discounting this by the 0.332 figure from moral weights (this includes sentience estimates, right?) gives a final DALY's per thousand of 1242 (or 1162 if we use the 3.5k figure directly) Under my laying period figures, the final estimate is 6352 Chicken-Dalys/thousand, which discounted by the RP moral weights comes to 2108 DALYs/thousand dollars. A similar analysis for broilers gives 1500 chicken-dalys per thousand dollars and 506 DALY's per thousand dollars. The default values from the cross cause website should match with either Laura's or mines estimates. 

Saulius is saying that each dollar affects 54 chicken years of life, equivalent to moving 54 chickens from caged to cage free environments for a year. The DALY conversion is saying that, in that year, each chicken will be 0.23 DALY’s better off. So in total, 54*0.23 = 12.43

I don't believe Saulius's numbers are directly used at any point in the model or intended to be used. The model replicates some of the work to get to those numbers. That said, I do think that you can use your approach to validate the model. I think the key discrepancy here is that the... (read more)

4
titotal
 Okay, I was looking at the field DALYs per bird per year" in this report, which is 0.2 matching with I have replicated. The 0.23 figure is actually something else, which explains a lot of the confusion in this conversation. I'll include my calculation at the end.  Before I continue, I want to thank you for being patient and working with me on this. I think people are making decisions based on these figures so it's important to be able to replicate them.  This report states that Saulius's numbers are being used: I think I've worked it out: if we take the 2.18 birds affected per year and multiply by the 15 year impact, we get 32.7 chicken years affected/dollar , which is the same as the 54 chicken years given by saulius discounted by 40% (54*0.6 = 32.4). This is the number that goes into the 0.23 figure, and this does already take into account the 15 years of impact.  So I don't get why there's still a discrepancy: although we take different routes to get there, we have the same numbers and should be getting the same results. My calculation, taken from here. laying time is 40 to 60 weeks, so we’ll assume it goes for exactly 1 year.  Disabiling: 430-156 = 274 hours disabling averted Hurtful = 4000-1741 = 2259 hours hurtful averted.  Annoying  6721-2076 =4645 hours annoying averted.  Total DALYs averted: 4.47*274/(365*24) = 0.14 disabling DALYS averted 0.15*2259/(365*24) = 0.0386 hurtful DALYS averted 0.015* 4645/(365*24) =0.00795hurtful Dalys averted Total is about 0.19 DALY’s averted per hen per year. 

If I take sallius's median result of 54 chicken years life affected per dollar, and then multiply by Laura's conversion number of 0.23 DALYs per $ per year, I get a result of 12.4 chicken years life affected per dollar. If I convert to DALY's per thousand dollars, this would result in a number of 12,420.

Laura’s numbers already take into account the number of chickens affected. The 0.23 figure is a total effect to all chickens covered per dollar per year. To get the effect per $1000, we need to multiply by the number of years the effect will last and by ... (read more)

2
titotal
I'm sorry, but this just isn't true. You can look at the field for "annual CC DALYs per bird per year" here (with the 0.2 value), it does not include Saulius's estimates. (I managed to replicate the value and checked it against the fields here, they match).  Saulius’s estimates already factor in the 14 year effect of the intervention. You’ll note that the “chickens affected per dollar” is multiplied by the mean years of impact when giving out the "12 to 160" result.   Saulius is saying that each dollar affects 54 chicken years of life, equivalent to moving 54 chickens from caged to cage free environments for a year. The DALY conversion is saying that, in that year, each chicken will be 0.23 DALY’s better off. So in total, 54*0.23 = 12.43 DALYs are averted per dollar, or 12430 DALYS per thousand, as I said in the last comment. However, I did notice in here that the result was deweighted by 20%-60% because they expected future campaigns to be less effective, which would bring it down to around 7458. I didn't factor in the moral conversions because those are seperate fields in the site. If I use P(sentience) of 0.8 and moral weight of 0.44 as the site defaults to, the final DALy per thousand should be 7458*0.8*0.44= 2386 DALYs/thousand dollars, about three times more than the default value on the site. 

I am understanding correctly that none of these factors are included in the global health and development effectiveness evaluation?

Correct!

A common response we see is that people reject the radical animal-friendly implications suggested by moral weights and infer that we must have something wrong about animals' capacity for suffering. While we acknowledge the limitations of our work, we generally think a more fruitful response for those who reject the implications is to look for other reasons to prefer helping humans beyond purely reducing suffering. (... (read more)

First, The google doc states that the life-years affected per dollar is 12 to 120, but Sallius report says it's range is 12 to 160. Why the difference? Is this just a typo in the google doc?

I believe that is a typo in the doc. The model linked from the doc uses a log normal distribution between 13 and 160 in the relevant row (Hen years / $). (I can't speak to why we chose 13 rather than 12, but this difference is negligible.)

Second, the default values in the tool are given as 160 to 3600. Why is this range higher (on a percentage basis) than the life years

... (read more)
5
titotal
Thanks, hope the typos will be fixed. I think I've almost worked through everything to replicate the results, but the default values still seem off. If I take sallius's median result of 54 chicken years life affected per dollar, and then multiply by Laura's conversion number of 0.23 DALYs per $ per year, I get a result of 12.4 chicken years life affected per dollar. If I convert to DALY's per thousand dollars, this would result in a number of 12,420.  This is outside the 90% confidence interval for the defaults given on the site, which gives it as "between 160 and 3.6K suffering-years per dollar". If I convert this to the default constant value, it gives the suggested value of 1,900, which is roughly ten time lower than the value if I take Sallius's median and laura's conversion factor.  If I put in the 12420 number into the field, the site gives out 4630 DALY's per thousand dollars, putting it about 10 times higher than originally stated in the post, which seems more in line with other RP claims (after all, right now the chicken campaign is presented as only 10 times more cost effective, whereas others are claiming it's 1000x more effective using RP numbers). 

That would require building in further assumptions, like a clip of the results at 100%. We would probably want to do that, but it struck me in thinking about this that it is easy to miss when working in a model like this. It is a bit counterintuitive that lowering the lower bound of a log normal distributions can increase the mean.

If I drop the lower bound by 4 orders of magnitude, to "between 0.0000002 and 0.87 times", I get a result of 709 Dalys/1000$, which is basically unchanged. Do sufficiently low bounds basically do nothing here?

This parameter is set to a normal distribution (which, unfortunately you can't control) and the normal distribution doesn't change much when you lower the lower bound. A normal distribution between 0.002 and 0.87 is about the same as a normal distribution between 0 and 0.87. (Incidentally, if the distribution were a lognormal distribution with the sam... (read more)

This parameter is set to a normal distribution (which, unfortunately you can't control) and the normal distribution doesn't change much when you lower the lower bound. A normal distribution between 0.002 and 0.87 is about the same as a normal distribution between 0 and 0.87. (Incidentally, if the distribution were a lognormal distribution with the same range, then the average result would fall halfway between the bounds in terms of orders of magnitude. This would mean cutting the lower bound would have a significant effect. However, the effect would actual

... (read more)
9
titotal
Thanks for clarifying! I think these numbers are the crux of the whole debate, so it's worth digging into them.  I am understanding correctly that none of these factors are included in the global health and development effectiveness evaluation? I'm not sure how this is different to my hypothetical, except in degree?   But the thing we are actually debating here is "should we prevent african children from dying of malaria, or prevent a lot of chickens from being confined to painful cages", which is an action. If you are using a weight of ~0.44 to make that decision, then shouldn't you similarly use it to make the "free 3 chickens or a human" decision? 

We have heard from some organizations that have taken a close look at the CCM and it has spawned some back and forth about the takeaways. I don't think I can disclose anything specific further at this point, though perhaps we might be able to in the future.

Thanks for reporting this. You found an issue that occurred when we converted data from years to hours and somehow overlooked the place in the code where that was generated. It is fixed now. The intended range is half a minute to 37 minutes, with a mean of a little under 10. I'm not entirely sure where the exact numbers for that parameter come from, since Laura Duffy produced that part of the model and has moved on to another org, but I believe it is inspired by this report. As you point out, that is less than three hours of disabling equivalent pain. I'll have to dig deeper to figure out the rationale here.

2
Vasco Grilo🔸
Thanks for the update, Derek. To give credit where it is due, it was Michael Johnston who found the issue.

After working on WIT, I’ve grown a lot more comfortable producing provisional answers to deep questions. In similar academic work, there are strong incentives to only try to answer questions in ways that are fully defensible: if there is some other way of going about it that gives a different result, you need to explain why your way is better. For giant nebulous questions, this means we will make very slow progress on finding a solution. Since these questions can be very important, it is better to come up with some imperfect answers rather than just workin... (read more)

One of the big prioritization changes I’ve taken away from our tools is within longtermism. Playing around with our Cross-Cause Cost-Effectiveness Model, it was clear to me that so much of the expected value of the long-term future comes from the direction we expect it to take, rather than just whether it happens at all. If you can shift that direction a little bit, it makes a huge difference to overall value. I no longer think that extinction risk work is the best kind of intervention if you’re worried about the long-term future. I tend to think that AI (non-safety) policy work is more impactful in expectation, if we worked through all of the details.

Thanks for raising this point. We think that choosing the right decision theory that can handle imprecise probabilities is a complex issue that has not been adequately resolved. We take the point that Mogensen’s conclusions have radical implications for the EA community at large and we haven’t formulated a compelling story about where Mogensen goes wrong. However, we also believe that there are likely to be solutions that will most likely avoid those radical implications, and so we don’t need to bracket all of the cause prioritization work until we find th... (read more)

Load more