All of Isaac Dunn's Comments + Replies

Isaac Dunn
5
7
0
50% agree

Before reading this quick take, how familiar were you with this forum’s voting guidelines?

I wasn't sure if I was, but reading the guidelines matched my guess of what they would say, so I think I was familiar with them. 

6
MichaelDickens
Funny, I thought the same thing but I voted on the opposite end of the spectrum. I suppose "how familiar are you with the voting guidelines?" is pretty open to interpretation.

Actually, computer science conferences are peer reviewed. They play a similar role as journals in other fields. I think it's just a historical curiosity that it's conferences rather than journals that are the prestigious places to publish in CS!

Of course, this doesn't change the overall picture of some AI work and much AI safety work not being peer reviewed.

Thanks, this back and forth is very helpful. I think I've got a clearer idea about what you're saying. 

I think I disagree that it's reasonable to assume that there will be a fixed N = 10^35 future lives, regardless of whether it ends up Malthusian. If it ends up not Malthusian, I think I'd expect the number of people in the future to be far less than whatever the max imposed by resource constraints is, ie much less than 10^35.

So I think that changes the calculation of E[saving one life], without much changing E[preventing extinction], because you need... (read more)

1
Avik Garg
Edit: I misread and thought you were saying non-Malthusian worlds had more lives at first; realized you said the opposite, so we're saying the same thing and we agree. Will have to do more math about this. This is an interesting point that I hadn't considered! I think you're right that non-Malthusian futures are much larger than Malthusian futures in some cases...though if i.e. the "Malthusian" constraint is digital lives or such, not sure. I think the argument you make actually cuts the other way. That to go back to the expected value...the case where the single death is deriving its EV from is precisely the non-Malthusian scenarios (when its progeny is not replaced by future progeny) so its EV actually remains the same. The extinction EV is the one that reduces...so you'll actually get a number much less than 10^10 if you have high credence that Malthusianism is true and think Malthusian worlds have more people. But, if you believe the opposite...that Malthusian worlds have more people, which I have not thought about but actually think might be true, yes a bigger gap than 10^10; will have to think about this. Thanks! Does this make sense to you?  

Ah nice, thanks for explaining! I'm not following all the calculations still, but that's on me, and I think they're probably right.

But I don't think your argument is actually that relevant to what we should do, even if it's right. That's because we don't care about how good our actions are as a fraction/multiple of what our other options are. Instead, we just want to do whatever leads to the best expected outcomes. 

Suppose there was a hypothetical world where there was a one in ten chance the total figure population was a billion, and 90% chance the p... (read more)

1
Avik Garg
Hmm, I’m not sure Iunderstand your point so maybe let me add some more numbers to what I’m saying and you could say if you think your point is responsive?  What I think you’re saying is that I’m estimating E[value saving one life / value stopping extinction] rather than E[value of saving one life] / E[value of stopping extinction]. I think this is wrong and that I’m doing the latter. I start from the premise of we want to save in expectation most lives (current and future are equivalent). Let’s say I have two options…I can prevent extinction or directly stop a random living person from dying. Assume there are 10^35 (I just want N >> C) future lives and there are 10^10 current lives. Now assume I believe there is a 99% chance that when I save this one life, fertility in the future somehow goes up such that the individual’s progeny are replaced, but there’s a 1% chance the individual’s progeny is not replaced. The individual is responsible for 10^35/10^10 =10^25 progeny. This gives E[stopping random living person from dying] ~ 1%*10^25 =10^23. And we’d agree E[preventing extinction] = 10^35. So E[value of saving one life] / E[value of stopping extinction] ~ 10^-12. Interestingly E[value of saving one life / value of stopping extinction] is the same in this case because the denominator is just a constant random variable…though E[value of stopping extinction/value of saving one life] is very very large (much larger than 10^12).

I think your calculations must be wrong somewhere, although I can't quite follow them well enough to see exactly where. 

If you have a 10% credence in Malthusianism, then the expected badness of extinction is 0.1*10^35, or whatever value you think a big future is. That's still a lot closer to 10^35 times the badness of one death than 10^10 times.

Does that seem right?

1
Avik Garg
No, because you have to compare the two harms.  Take the number of future lives as N and current population as C Extinction is as bad as N lives lost. One death is w/ 10% credence only approx as bad as 1 death bc Malthusianism. But w/ 90% credence, it is as bad as N/C lives lost. So, plugging in 10^35 as N and 10^10 as C, EV of one death is 1 (.1) + N/C (0.9) ~ N/C * 0.9 ~ 9e24, 11 times worse than extinction. In general, if you have credence p, extinction becomes 10^10*1/(1-p) worse than one death.

Agree coin flip is unacceptable! Or even much less than coin flip is still unacceptable.

I agree with this comment, but I interpreted your original comment as implying a much greater degree of certainty of extinction assuming ASI is developed than you might have intended. My disagree vote was meant to disagree with the implication that it's near certain. If you think it's not near certain it'd cause extinction or equivalent, then it does seem worth considering who might end up controlling ASI!

3
Davidmanheim
If it's "only" a coinflip if it causes extinction if developed today, to be wildly optimistic, then I will again argue that talking about who should flip the coin seems bad - the correct answer in that case is no one, and we should be incredibly clear on that!

You're stating it as a fact that "it is" a game of chicken, i.e. that it's certain or very likely that developing ASI will cause a global catastrophe because of misaligned takeover. It's an outcome I'm worried about, but it's far from certain, as I see it. And if it's not certain, then it is worth considering what people would do with aligned AI.

5
Davidmanheim
I'm confused why people think certainty is needed to characterize this as a game of chicken! It's certainly not needed in order for the game theoretic dynamics to apply. I can make a decision about whether to oppose something given that there is substantial uncertainty, and I have done so.

I heard reports of it getting out of sync or being out of date in some way. For example, a room change on Swapcard not being reflected in the Google calendar. I haven't tried it myself, and I haven't heard anything less vague, sorry. 

I think that the Google calendar syncing is at least a bit buggy for now, FYI. Agree good news though!

Ah interesting, good to know! What kind of bugs have you encountered? I did some basic tests and it seemed to work smoothly for me.

7
calebp
Not sure right now, but probably sometime next quarter.

Thanks Vasco! :)

I agree that thinking about other moral theories is useful for working out what utilitarianism would actually recommend.

That's an interesting point re increasing the total amount of killing, I hadn't considered that! But I was actually picking up on your comment which seemed to say something more general - that you wouldn't intrinsically take into account whether an option involved (you) killing people, you'd just look at the consequences (and killing can lead to worse consequences, including in indirect ways, of course). But it sounds like... (read more)

2
Vasco Grilo🔸
Yes.

Do you not worry about moral uncertainty? Unless you're certain about consequentialism, surely you should put some weight on avoiding killing even if it maximises impartial welfare?

3
Vasco Grilo🔸
Hi Isaac. I fully endorse expected total hedonistic utilitarianism (ETHU) in principle. However, I think it is often good to think about the implications of other moral theories as heuristics to follow ETHU well in practice. I think saving human lives increases the number of beings killed via increasing the number of farmed and wild animals killed.

You're welcome! N=1 though, so might be worth seeing what other people think too.

For what it's worth, although I do think we are clueless about the long-run (and so overall) consequences of our actions, the example you've given isn't intuitively compelling to me. My intuition wants to say that it's quite possible that the cat vs dog decision ends up being irrelevant for the far future / ends up being washed out.

Sorry, I know that's probably not what you want to hear! Maybe different people have different intuitions.

3
Jim Buhler
That's very useful, thanks! I was hoping that it felt like there is no way it gets washed out given that what is such a large portion of the World's resources gets put into this, so really good to know you don't have this intuition reading this (especially if you generally think we are clueless!).  Maybe I can give a better intuition pump for how the effects will last and ramificate. But, also, maybe talking about cats and dogs makes the decision look too trivial to begin with and other cause area examples would be better. Thanks again! Glad you shared an intuition that goes against what I was hoping. That was the whole point of me posting this :)

I don't think OpenAI's near term ability to make money (e.g. because of the quality of its models) is particularly relevant now to its valuation. It's possible it won't be in the lead in the future, but I think OpenAI investors are betting on worlds where OpenAI does clearly "win", and the stickiness of its customers in other worlds doesn't really affect the valuation much.

So I don't agree that working on this would be useful compared with things that contribute to safety more directly.

How much do you think customers having 0 friction to switching away fro... (read more)

1
Ebenezer Dukakis
They're losing billions every year, and they need a continuous flow of investment to pay the bills. Even if current OpenAI investors are focused on an extreme upside scenario, that doesn't mean they want unlimited exposure to OpenAI in their portfolio. Eventually OpenAI will find themselves talking to investors who care about moats, industry structure, profit and loss, etc. The very fact that OpenAI has been throwing around revenue projections for the next 5 years suggests that investors care about those numbers. I also think the extreme upside is not that compelling for OpenAI, due to their weird legal structure with capped profit and so on? On the EA Forum it's common to think in terms of clear "wins", but it's unclear to me that typical AI investors are thinking this way. E.g. if they were, I would expect them to be more concerned about doom, and OpenAI's profit cap. Dario Amodei's recent post was rather far out, and even in his fairly wild scenario, no clear "win" was implied or required. There's nothing in his post that implies LLM providers must be making outsized profits -- same way the fact that we're having this discussion online doesn't imply that typical dot-com bubble companies or telecom companies made outsized profits. If it becomes common knowledge that LLMs are bad businesses, and investor interest dries up, that could make the difference between OpenAI joining the ranks of FAANG at a $1T+ valuation vs raising a down round. Markets are ruled by fear and greed. Too much doomer discourse inadvertently fuels "greed" sentiment by focusing on rapid capability gain scenarios. Arguably, doomer messaging to AI investors should be more like: "If OpenAI succeeds, you'll die. If it fails, you'll lose your shirt. Not a good bet either way." There are liable to be tipping points here -- chipping in to keep OpenAI afloat is less attractive if future investors are seeming less willing to do this. There's also the background risk of a random recession due to

I think investors want to invest in OpenAI so badly almost entirely because it's a bet on OpenAI having better models in the future, not because of sticky customers. So it seems that the effect of this on OpenAI's cost of capital would be very small?

2
Ebenezer Dukakis
OpenAI models will improve, and offerings from competitors will also improve. But will OpenAI's offerings consistently maintain a lead over competitors? Here is an animation I found of LLM leaderboard rankings over time. It seems like OpenAI has consistently been in the lead, but its lead tends to be pretty narrow. They might even lose their lead in the future, given the recent talent exodus. [Edit: On the other hand, it's possible their best models are not publicly available.] If switching costs were zero, it's easy for me to imagine businesses becoming price-sensitive. Imagine calling a wrapper API which automatically selects the cheapest LLM that (a) passes your test suite and (b) has a sufficiently low rate of confabulations/misbehavior/etc. Given the choice of an expensive LLM with 112 IQ, and a cheap LLM with 110 IQ, a rational business might only pay for the 112 IQ LLM if they really need those additional 2 IQ points. Perhaps only a small fraction of business applications will fall in the narrow range where they can be done with 112 IQ but not 110 IQ. For other applications, you get commoditization. A wrapper API might also employ some sort of router model that tries to figure out if it's worth paying extra for 2 more IQ points on a query-specific basis. For example, initially route to the cheapest LLM, and prompt that LLM really well, so it's good at complaining if it can't do the task. If it complains, retry with a more powerful LLM. If the wrapper API was good enough, and everyone was using it, I could imagine a situation where even if your models consistently maintain a narrow lead, you barely eke out extra profits. It's possible that https://openrouter.ai/ is already pretty close to what I'm describing. Maybe working there would be a good EA job?

Interesting exercise, thanks! The link to view the questions doesn't work though. It says:

The form AI Grantmaking Priorities Survey is no longer accepting responses.
Try contacting the owner of the form if you think that this is a mistake.

Interesting!

I think my worry is people who don't think they need advice about what the future should look like. When I imagine them making the bad decision despite having lots of time to consult superintelligent AIs, I imagine them just not being that interested in making the "right" decision? And therefore their advisors not being proactive in telling them things that are only relevant for making the "right" decision.

That is, assuming the AIs are intent aligned, they'll only help you in the ways you want to be helped:

  • Thoughtful people might realise the im
... (read more)
3
Zach Stein-Perlman
Briefly + roughly (not precise): At some point we'll send out lightspeed probes to tile the universe with some flavor of computronium. The key question (for scope-sensitive altruists) is what that computronium will compute. Will an unwise agent or incoherent egregore answer that question thoughtlessly? I intuit no. I can't easily make this intuition legible. (So I likely won't reply to messages about this.)

I agree that the text an LLM outputs shouldn't be thought of as communicating with the LLM "behind the mask" itself.

But I don't agree that it's impossible in principle to say anything about the welfare of a sentient AI. Could we not develop some guesses about AI welfare by getting a much better understanding of animal welfare? (For example, we might learn much more about when brains are suffering, and this could be suggestive of what to look for in artificial neural nets)

It's also not completely clear to me what the relationship between the sentient being ... (read more)

2
LGS
I should not have said it's in principle impossible to say anything about the welfare of LLMs, since that too strong a statement. Still, we are very far from being able to say such a thing; our understanding of animal welfare is laughably bad, and animal brains don't look anything like the neural networks of LLMs. Maybe there would be something to say in 100 years (or post-singularity, whichever comes first), but there's nothing interesting to say in the near future. This is a weird EA-only intuition that is not really shared by the rest of the world, and I worry about whether cultural forces (or "groupthink") are involved in this conclusion. I don't know whether the total amount of suffering is more than the total amount of pleasure, but it is worth noting that the revealed preference of living things is nearly always to live. The suffering is immense, but so is the joy; EAs sometimes sound depressed to me when they say most life is not worth living. To extrapolate from the dubious "most life is not worth living" to "LLMs' experience is also net bad" strikes me as an extremely depressed mentality, and one that reminds me of Tomasik's "let's destroy the universe" conclusion. I concede that logically this could be correct! I just think the evidence is so weak is says more about the speaker than about LLMs.

Why does "lock-in" seem so unlikely to you?

One story:

  • Assume AI welfare matters
  • Aligned AI concentrates power in a small group of humans
  • AI technology allows them to dictate aspects of the future / cause some "lock in" if they want. That's because:
    • These humans control the AI systems that have all the hard power in the world
    • Those AI systems will retain all the hard power indefinitely; their wishes cannot be subverted
    • Those AI systems will continue to obey whatever instructions they are given indefinitely
  • Those humans decide to dictate some or all of what the fut
... (read more)
3
Zach Stein-Perlman
I agree this is possible, and I think a decent fraction of the value of "AI welfare" work comes from stuff like this. This would be very weird: it requires that either the value-setters are very rushed or that they have lots of time to consult with superintelligent advisors but still make the wrong choice. Both paths seem unlikely.

Good question! I share that intuition that preventing harm is a really good thing to do, and I find striking the right balance between self-sacrifice and pursuing my own interests difficult.

I think if you argue that that leads to anything close to a normal life you are being disingenuous

I think this is probably wrong for most people. If you make yourself unhappy by trying to force yourself to make sacrifices you don't want to make, I think most people will be much less productive. And I think that most people actually need a fairly normal social life etc. ... (read more)

I think misaligned AI values should be expected to be worse than human values, because it's not clear that misaligned AI systems would care about eg their own welfare.

Inasmuch as we expect misaligned AI systems to be conscious (or whatever we need to care about them) and also to be good at looking after their own interests, I agree that it's not clear from a total utilitarian perspective that the outcome would be bad.

But the "values" of a misaligned AI system could be pretty arbitrary, so I don't think we should expect that.

This is a true, counterfactual match, and we will only receive the equivalent amount to what we can raise.

What will happen to the money counterfactually? Presumably it will be donated to other things the match funder thinks are roughly as good as GWWC?

6
Luke Freeman 🔸
Thanks both. They haven't shared this with us specifically so I can't speak for them. They have been very clear that it is a conditional match. I'll try updating the wording for clarity.

I'm also confused by this. The use of "and" (instead of, say, "in that", "because", or "to the extent that") suggests that they've verified counterfactuality in some stronger way than just "the money won't go to us this season if you don't donate", but then they should be telling us how they know this.

Is this a problem? Seems fine to me, because the meaning is often clear, as in two of your examples, and I think it adds value in those contexts. And if it's not clear, doesn't seem like a big loss compared to a counterfactual of having none of these types of vote available.

I think that trying to get safe concrete demonstrations of risk by doing research seems well worth pursuing (I don't think you were saying it's not).

Do you have any thoughts on how should people decide between working on groups at CEA and running a group on the ground themselves?

I imagine a lot of people considering applying could be asking themselves that question, and it doesn't seem obvious to me how to decide.

Hi Isaac, this is a good question! I can elaborate more in the Q&A tomorrow but here are some thoughts:

Ultimatley a lot depends on your personal fit and comparative advantage. I think people should do the things they excel at. While I do think you can have a more scalable impact on the groups team, the groups team would have very little to no impact without the organizers working on the ground! 

I can share some of the reasons that led me to prefer working at CEA over working on the ground:

  • I value having close management to help me think through my
... (read more)

To be fair, I think I'm partly making wrong assumptions about what exactly you're arguing for here.

On a slightly closer read, you don't actually argue in this piece that it's as high as 90% - I assumed that because I think you've argued for that previously, and I think that's what "high" p(doom) normally means.

2
Greg_Colbourn ⏸️
I do think it is basically ~90%, but I'm arguing here for doom being the default outcome of AGI; I think "high" can reasonably be interpreted as >50%.

Relatedly, I also think that your arguments for "p(doom|AGI)" being high aren't convincing to people that don't share your intuitions, and it looks like you're relying on those (imo weak) arguments, when actually you don't need to

4
Greg_Colbourn ⏸️
I'm crying out for convincing gears-level arguments against (even have $1000 bounty on it), please provide some.

I think you come across as over-confident, not alarmist, and I think it hurts how you come across quite a lot. (We've talked a bit about the object level before.) I'd agree with John's suggested approach.

-1
William the Kiwi
What part of Greg writing comes across as over confident?
2
Greg_Colbourn ⏸️
I feel like this is a case of death by epistemic modesty, especially when it isn't clear how these low p(doom) estimates are arrived at in a technical sense (and a lot seems to me like a kind of "respectability heuristic" cascade). We didn't do very well with Covid as a society in the UK (and many other countries), following this kind of thinking.
5
Isaac Dunn
Relatedly, I also think that your arguments for "p(doom|AGI)" being high aren't convincing to people that don't share your intuitions, and it looks like you're relying on those (imo weak) arguments, when actually you don't need to

Makes sense. To be clear, I think global health is very important, and I think it's a great thing to devote one's life to! I don't think it should be underestimated how big a difference you can make improving the world now, and I admire people who focus on making that happen. It just happens that I'm concerned the future might be even higher priority thing that many people could be in a good position to address.

On your last point, if you believe that the EV from a "effective neartermism -> effective longtermism" career change is greater than a "somewhat harmful career -> effective neartermism" career change, then the downside of using a "somewhat harmful career -> effective longtermism" example is that people might think the "stopped doing harm" part is more important than the "focused on longtermism" part.

More generally, I think your "arguments for the status quo" seem right to me! I think it's great that you're thinking clearly about the considerations on both sides, and my guess is that you and I would just weight these considerations differently.

7
NickLaing
Thanks Isaac for the encouragement on the validity of the "arguments for the staus quo". I'm not really sure how I weight the considerations to be honest, I'm more raising the questions for discussion. Yes that's a fair point about the story at the end. I hadn't consider the "stopped doing harm" part might make my example confusiton. Maybe then I would prefer a "went from being a doctor" to "focused on longtermism", because otherwise it feels like a bit of a kick in the teeth to a decent chunk of the EA community who have decided that global health is.a great thing to devote your life to ;).

Thank you for sharing these! I'm probably going to try the first three as a result of this post.

Another thing on my mind is that we should beware surprising and suspicious convergence - it would be surprising and suspicious if the same intervention (present-focused WAW work) was best for improving animals' lives today and also happened to be best for improving animals' lives in the distant future.

I worry about people interested in animal welfare justifying maintaining their existing work when they switch their focus to longtermism, when actually it would be better if they worked on something different.

3
Jens Aslaug 🔸
(I hope it’s not confusing that I'm answering both your comments at once).  While I will have to consider this for longer, my preliminary thought is that I agree with most of what you said. Which means that I might not believe in some of my previous statements.   Thanks for the link to that post. I do agree and I can definitely see how some of these biases have influenced a couple of my thoughts.  -- Okay, I see. Well actually, my initial thought was that all of those four options had a similar impact on the longterm future. Which would justify focusing on short-term interventions and advocacy (which would correspond with working on point number three and four). However after further consideration, I think the first two are of higher impact when considering the far future. Which means I (at least for right now) agree with your earlier statement:  While I still think the “flow through effect” is very real for WAW, I do think that it’s probably true working on s-risks more directly might be of higher impact.  -- I was curious if you have some thoughts on these conclusions (concluded based on a number of things you said and my personal values):  * Since working on s-risk directly is more impactful than working on it indirectly, direct work should be done when possible.  * There is no current organization working purely on animal related s-risk (as far as I know). So if that’s your main concern, your options are start-up or convincing an “s-risk mitigation organization” that you should work on this area full time. * Animal Ethics works on advocating moral circle expansion. But since this is of less direct impact to the longterm future, this has less of an effect on reducing s-risk than more direct work.  * If you’re also interested in reducing other s-risks (e.g. artificial sentience), then working for an organization that directly tries to reduce the probability of a number of s-risk is your best option (e.g. Center on Long-Term Risk or Center for Reducin

Thanks for your reply! I can see your perspective.

On your last point, but future-focused WAW interventions, I'm thinking of things that you mention in the tractability section of your post:

Here is a list of ways we could work on this issue (directly copied from the post by saulius[9]):

“To reduce the probability of humans spreading of wildlife in a way that causes a lot of suffering, we could:

  1. Directly argue about caring about WAW if humans ever spread wildlife beyond Earth
  2. Lobby to expand the application of an existing international law that tries to protect
... (read more)

For the kinds of reasons you give, I think it could be good to get people to care about the suffering of wild animals (and other sentient beings) in the event that we colonise the stars.

I think that the interventions that decrease the chance of future wild animal suffering are only a subset of all WAW things you could do, though. For example, figuring out ways to make wild animals suffer less in the present would come under "WAW", but I wouldn't expect to make any difference to the more distant future. That's because if we care about wild animals, we'll fi... (read more)

4
Jens Aslaug 🔸
I do agree that current WAW interventions have a relatively low expected impact compared with other WAW work (e.g. moral circle expansion) if only direct effects are counted.  Here are some reasons why I think current interventions/research may help the longterm future.  * Doing more foundational work now means we can earlier start more important research and interventions, when the technology is available. (Probably a less important factor) * Current research gives us a better answer to how much pleasure and suffering wild animals experience, which helps inform future decisions on the spread of wildlife. (This may not be that relevant yet) * Showcasing that interventions can have a positive effect on the welfare of wildlife, could help convince more people that helping wildlife is tractable and the morally right thing to do (even if it’s unnatural). (I think this to be the most important effect)  So I think current interventions could have a significant impact on moral circle expansion. Especially because I think you have to have two beliefs to care for WAW work: believe that the welfare of wildlife is important (especially for smaller animals like insects, which likely make up the majority of suffering) and believe that interfering with nature could be positive for welfare. The latter may be difficult to achieve without proven interventions since few people think we should intervene in nature.  Whether direct moral circle expansion or indirect (via. interventions) are more impactful are unclear to me. Animal Ethics mainly work on the former and Wild Animal Initiative works mainly on the latter. I’m currently expecting to donate to both.   I think having an organization working directly on this area could be of high importance (as I know only the Center For Reducing Suffering and the Center on Long-Term Risk work partly on this area). But how do you think it’s possible to currently work on "future-focused wild animal welfare interventions"? Other than doing

If I understand correctly, you put 0.01% on artificial sentience in the future. That seems overconfident to me - why are you so certain it won't happen?

3
Jens Aslaug 🔸
Yes that’s correct and I do agree with you. To be honest the main reasons were due to limited knowledge and simplification reasons. Putting any high number for the likelihood of “artificial sentience” would make it the most important cause area (which based on my mindset, it might be).  But I’m currently trying to figure out which of the following I think is the most impactful to work on: AI-alignment, WAW or AI sentience. This post was simply only about the first two.  When all of that’s been said, I do think AI-sentience is a lot less likely than many EAs think (which still doesn't justify “0.01%”). But note that this is just initial thoughts based on limited information. Anyways, here’s my reasoning:  * While I do agree that it might be theoretically possible and could cause suffering on an astronomical scale, I do not understand why we would intentionally or unintentionally create it. Intentionally I don't see any reason why a sentient AI would perform any better than a non-sentient AI. And unintentionally, I could imagine that with some unknown future technology, it might be possible. But no matter how complex we make AI with our current technology, it will just become a more "intelligent" binary system.  * Even if we create it, it would only be relevant as an s-risk if we don’t realize it and fix it.  However I do think the probability of me changing my mind is high. 

I've only skimmed this, but just want to say I think it's awesome that you're doing your own thinking trying to compare these two approaches! In my view, you don't need to be "qualified" to try to form your own view, which depends on understanding the kinds of considerations you raise. This decision matters a lot, and I'm glad you're thinking carefully about it and sharing your thoughts.

4
Jens Aslaug 🔸
Thank you for your encouraging words! I appreciate your support and perspective.

I interpreted the title of this post as a bill banning autonomous AI systems from paying people to do things! I did think it was slightly early.

0[anonymous]
Actually, I think I might have done that. (I say this so you don't feel gaslighted by the change in name.)
0[anonymous]
How should I change the name to prevent this?

Would you be eligible for the graduate visa? https://www.gov.uk/graduate-visa

If so, would that meet your needs?

2
𝕮𝖎𝖓𝖊𝖗𝖆
Thanks, yeah. My main hesitancy about this is that I probably want to go for a PhD, but can only get the graduate visa once, and I may want to use it after completing the PhD. But I've come around to maybe it being better to use it up now, pursue a PhD afterwards, and try to secure employment before completing my program so I can transfer to the skilled workers visa.

(I've just realised this is close to just a rephrasing of some of the other suggestions. Could be a helpful rephrasing though.)

The Superalignment team's goal is "to build a roughly human-level automated alignment researcher".

Human-level AI systems sound capable enough to cause a global catastrophe if misaligned. So is the plan to make sure that these systems are definitely aligned (if so, how?), or to make sure that they are deployed in a such a way that they would not be able to take catastrophic actions even if they want to (if so, what would that look like?)?

1
Isaac Dunn
(I've just realised this is close to just a rephrasing of some of the other suggestions. Could be a helpful rephrasing though.)

Thanks David, that's just the kind of reply I was hoping for! Those three goals do seem to me like three of the most important. It might be worth adding that context to your write-up.

I'm curious whether there's much you did specifically to achieve your third goal - inspiring people to take action based on high quality reasoning - beyond just running an event where people might talk to others who are doing that. I wouldn't expect so, but I'd be interested there was.

3
David M
We did encourage speakers to include action points and action-relevant information in their content, and tried to prioritise action-relevant workshops (e.g. "what it takes to found a charity"); I think that's about all. Thanks for the tip to include the goals in the write-up.

Thanks for writing this up! I'd be interested if you had time to say more about what you think the main theory of change of the event was (or should have been).

5
David M
What I'll say should be taken more as representative of how I've been thinking, than of how CEA or other people think about it. These were our objectives, in order: 1: Connect the EA UK community. 2: Welcome and integrate less well-connected members of the community. Reduce the social distance within the UK EA community. 3: Inspire people to take action based on high-quality reasoning. The main emphasis was on 1, where the theory of impact is something like: The EA community will achieve more by working together than they will by working as individuals; facilitating people to build connections makes collaboration more likely. Some valuable kinds of connections might be: mentoring relationships, coworking, cofounding, research collabs, and not least friendships (for keeping up one's motivation to do good). We added other goals beyond connecting people, since a lot of changes to plans will come from one-off interactions (or even exposures to content); I think of someone deciding to apply for funding after attending a workshop on how to do that. Plausibly though, longer-lasting, deeper connections dominate the calculation, because of the 'heavy tail' of deep collaborations, such as an intern hire I heard of which resulted from this conference. I'll tag @OllieBase (CEA Events) in case he wants to give his own answer to this question.

Interesting results, thanks for sharing! I think getting data from people who attend events is an important source of information about what's working and what's not.

I do worry a bit about what's best for the world coming apart from what people report as being valuable to them. (This comment ended up a bit rambley, sorry.)

Two main reasons that might be the case:

  1. If the event causes someone's goals or motivations to change in a way that's great for the world, my guess is that doesn't feel valuable to the person compared to helping the person get or do things
... (read more)
2
OllieBase
Thanks Isaac, I agree relying on self-report is a key limitation here. In fact, when reviewing people's stories I would often wish they expanded on something that seemed small to them but important to us (e.g. they'd write "became more interested in pursuing X path" as part of a list, but that stood out to me as something exciting from an impact perspective). I didn't mention this in the report, but I also do user interviews fairly regularly to get some more colour on things like this and followed up with several people whose stories seemed impactful. I wouldn't say the events team are following the baseline of "let's do things that people report as valuable", and are just using that as one guiding light (albeit a significant one). I agree forming a clearer framework of how people arrive at impactful work would be exciting.

Are there any lessons that GWWC has learnt that you think would be useful for EA community builders to know and remember?

8
Luke Freeman 🔸
Thanks Isaac! Personal advocacy goes a long way, especially when you are kind, warm, meet people where they are at and inspire them to see where they could be. Sometimes it takes a long time to pay off but if you leave people intrigued and inspired it often comes back around. Most pledges and new donors can be traced back to some form of personal advocacy. Secondly, giving people specific calls to action (e.g. if you care about X you can donate to Y) is a very empowering move that leaves people feeling like they have agency and makes it easier for them to advocate for the ideas you present to them (whereas "look at how big and hard and complicated the world's problems are, this specific one is most important, and we need a very unique specific person to solve it" is going to be really demoralising and makes it ultimately harder to find those specific people anyway). 

If GWWC goes very well over the next five years (say 90th percentile), what would that look like?

Do you think that most of GWWC's impact will come from money moved, or from introducing people to EA who then change their career paths, or something else? (I can't tell immediately tell from your strategy, which mentions both.)

6
Luke Freeman 🔸
Thanks Isaac, good question! I believe that although donations moved through our platform are significant and a robust measurement, they only represent a part of GWWC's overall impact. In fact, over half the donations that we're aware of (via member reporting) are made directly to charities or via other donation platforms. In my view, if we truly make strides towards achieving our mission, I expect that the indirect impact will dwarf the direct impact and more easily measured impact. This could take various forms such as introducing people to EA principles and causes, promoting positive values, and influencing donations more loosely, and shifting the impact-orientation of the philanthropic sector, to name a just few. On a day to day basis however, our approach is to focus on what we can effectively measure and optimise, which predominantly includes donations and number of people giving effectively. However, we always have our broader mission in sight to ensure we don't make decisions that compromise greater impact elsewhere. This post by Joey gives examples of ways that organisations could undermine overall impact by focusing too much on just their own metrics. Essentially, we are accountable to our mission more than just the easily measurable metrics precisely because we think that's what's more important. If you ask anyone at the GWWC team you'd likely hear that they're frequently asked questions like "is this moving us towards our mission?" or "could this harm our overall mission?". We're planning to incorporate more indirect impact measures in our next impact evaluation, starting with the ones easier to measure like organisations using our research, or people we've referred to other organisations as donors or employees etc. Thanks again for your question, Isaac!

What is the best reason to think that GWWC isn't good for the world, in your view?

0
Luke Freeman 🔸
The two most plausible ones to me are that: 1. If our team could be having more impact working on other things and that us existing and them being currently bought into the case for impact is the reason they don't. 2. We poison the well/turn people off/create misconceptions etc we don't live up to our standards (e.g. quality and tone). That being said I think that generally speaking GWWC is one of the lower downside risk options within the EA community (especially compared with some of the (potential or actual) longtermist projects which I often worry much more about potential downsides or accidentally doing harm – in the active sense as opposed to the opportunity cost harm). If you have some in mind yourself I'd happy hear them out 😀 

Even if it's true that it can be hard to agree or disagree with a post as a whole, I do get the impression that people sometimes feel like they disagree with posts as a whole, and so simply downvote the post.

Also, I suspect it is possible to disagree with a post as a whole. Many posts are structured like "argument 1, argument 2, argument 3, therefore conclusion". If you disagree with the conclusion, I think it's reasonable to say that that's disagreeing with the post as a whole. If you agree with the arguments and the conclusion, then you agree with the po... (read more)

Load more