All of Aryeh Englander's Comments + Replies

[Disclaimers: My wife Deena works with Kat as a business coach - see my wife's comment elsewhere on this post. I briefly met Kat and Emerson while visiting in Puerto Rico and had positive interactions with them. My personality is such that I have a very strong inclination to try to see the good in others, which I am aware can bias my views.]

A few random thoughts related to this post:

1. I appreciate the concerns over potential for personal retaliation, and the other factors mentioned by @Habryka and others for why it might be good to not delay this kind of ... (read more)

This is great! Just wanted to mention that this kind of weighting approach works very well with the recent post A Model-based Approach to AI Existential Risk, by Sammy Martin, Lonnie Chrisman, and myself, particularly the section on Combining inside and outside view arguments. Excited to see more work in this area!

Looking over that comment, I realize I don't think I've seen anybody else use the term "secret sauce theory", but I like it. We should totally use that term going forward. :)

Fair. I suppose there are actually two paths to being a doomer (usually): secret sauce theory or extremely short timelines.

2
Greg_Colbourn
10mo
I'm not sure if the secret sauce adds anything re doomerism. Many non-doomers are arguably non-doomers because they think the secret sauce makes the AI humanlike enough that things will be fine by default - the AI will automatically be moral, "do the right thing", "see reason", or "clearly something intelligent would realise killing everyone to make paperclips is stupid" or something (and I think this kind of not-applying-the-Copernican-Revolution-to-mindspace is really dangerous).
3
Aryeh Englander
10mo
Looking over that comment, I realize I don't think I've seen anybody else use the term "secret sauce theory", but I like it. We should totally use that term going forward. :)

Many people who worry about AI x-risk believe some variation of this.

I've been meaning to ask: Are there plans to turn your Cold Takes posts on AI safety and The Most Important Century into a published book? I think the posts would make for a very compelling book, and a book could reach a much broader audience and would likely get much more attention. (This has pros and cons of course, as you've discussed in your posts.)

As I mentioned on one of those Facebook threads: At least don't bill the event as a global conference for EA people and then tell people no you can't come. Call it maybe the EA Professionals Networking Event or something, which (a) makes it clear this is for networking and not the kind of academic conference people might be used to, and (b) implies this might be exclusive. But if you bill it as a global conference, then make it be like a global conference. And at the very least make it very clear that it's exclusive! Personally I didn't notice any mention of exclusivity at all in any EA Global posts or advertising until I heard about people actually getting rejected and feeling bad about that.

My interpretation of the “Global” part in EAG is ‘from around the world’, not ‘everyone is invited’. E.g. for EAGxAustralia it seems like you’re much more likely to get accepted if you’re based in Australia or the Asia Pacific, because it’s about building the community there. But EA Global is about connecting people across these different communities, and doesn’t prioritise admissions based on geographical closeness.

Honestly I’m super confused why people perceive ‘EA Global’ as an inclusive-sounding name. Especially in contrast to ‘EAGx’, which evokes the TEDx vs TED contrast, where TEDxes have a much lower bar, are scrappier and more community based.

4
Guy Raveh
2y
I mean, from the moment I first had to apply I felt it was exclusive - but only in the sense of "they select the people whom the conference would most help to increase their impact". So once I was rejected from one, I didn't feel offended. I did get accepted for the very first one I applied to, so maybe that's the reason. I was already a few years in the movement by then though.
4
Locke
2y
What are these FB threads that are being referenced out of curiosity?  Feels very in-groupey.

Here's a perspective I mentioned recently to someone:

Many people in EA seem to think that very few people outside the "self identifies as an EA" crowd really care about EA concerns. Similarly, many people seem to think that very few researchers outside of a handful of EA-affiliated AI safety researchers really care about existential risks from AI.

Whereas my perspective tends to be that the basic claims of EA are actually pretty uncontroversial. I've mentioned some of the basic ideas many times to people and I remember getting pushback I think only once - a... (read more)

Good point! I started reading those a while ago but got distracted and never got back to them. I'll try looking at them again.

In some cases it might be easier to do this as a structured interview rather than asking for written analyses. For example, I could imagine that it might be possible to create a podcast where guests are given an article or two to read before the interview, and then the interviewer asks them for their responses on a point-by-point basis. This would also allow the interviewer (if they're particularly good) to do follow-up questions as necessary. On the other hand, my personal impression is that written analyses tend to be more carefully argued and thought through than real-time interviews.

3
alex lawsen (previously alexrjl)
2y
Have you read the transcripts from Vael Gates's structured interviews with people working in AI about safety stuff, they seem to have done something pretty close to what you're asking.

Thought: In what ways do EA orgs / funds go about things differently than in the rest of the non-profit (or even for-profit) world? If they do things differently: Why? How much has that been analyzed? How much have they looked into the literature / existing alternative approaches / talked to domain experts?

Naively, if the the thing they do differently is not related to the core differences between EA / that org and the rest of the world, then I'd expect that this is kind of like trying to re-invent the wheel and it won't be a good use of resources unless you have a good reason to think you can do better.

Thank you for posting this! I was going to post something about this myself soon, but you beat me to it!

Decision Analysis (the practical discipline of analyzing decisions, usually in a business, operations, or policy context; not the same as decision theory): This discipline overlaps in obvious ways with a lot of EA and LessWrong discussions, but I have seen few direct references to Decision Analysis literature, and there seems to be little direct interaction between the EA/LW and DA communities. I'd love to see if we could bring in a few DA experts to giv... (read more)

7
Jan-Willem
2y
Hi Aryeh, really interested in this as well. Can you link me to any literature, experts, videos, software etc. that you deem valuable from DA? Would be really useful for future training programs from Training For Good!
2
emwalz
2y
Which DA software systems are worth looking closer at? Have you done any preliminary analysis to know? I would be keen to find one worth using…

Some - see the links at the end of the post.

1
Michael_Wiebe
2y
Would be really helpful to have this front and center.

What I do (assuming I get to that point in the conversation) is that I deliberately mention points like this, even before trying to argue otherwise. In my experience (which again is just my experience) a good portion of the time the people I'm talking to debunk those counterarguments themselves. And if they don't, well then I can start discussing it at that point - but at that point it feels to me like I've already established credibility and non-craziness by (a) starting off with noncontroversial topics, (b) starting off the more controversial topics with... (read more)

I haven't read most of the post yet, but already I want to give a strong upvote for (1) funding critiques of EA, and (2) the fact that you are putting up a list of projects you'd like to see. I would like to see more lists of this type! I've been planning to do one of them myself, but I haven't gotten to it yet.

I think I mostly lean towards general agreement with this take, but with several caveats as noted by others.

On the one hand, there are clearly important distinctions to be made between actual AI risk scenarios and Terminator scenarios. On the other hand, in my experience people pattern-matching to the Terminator usually doesn't make anything seem less plausible to them, at least as far as I could tell. Most people don't seem to have any trouble separating the time travel and humanoid robot parts from the core concern of misaligned AI, especially if you imm... (read more)

Yes, I have seen people become more actively interested in joining or promoting projects related to AI safety. More importantly, I think it creates an AI safety culture and mentality. I'll have a lot more to say about all of this in my (hopefully) forthcoming post on why I think promoting near-term research is valuable.

3
aogara
2y
Strongly agreed that working on the near term applications of AI safety is underrated by most EAs. Nearly all of the AI safety discussion focuses on advanced RL agents that are not widely deployed in the world today, and it’s possible that these systems do not soon reach commercial viability. Misaligned AI is causing real harms today and solving those problems would be a great step towards building the technical tools and engineering culture necessary to scale to aligning more advanced AI. (That’s just a three sentence explanation of a topic deserving much more detailed analysis, so really looking forward to your post!!)

[Disclaimer: I haven't read the whole post in detail yet, or all the other comments, so apologies if this is mentioned elsewhere. I did see that the Partnerships section talks about something similar, but I'm not sure it's exactly what I'm referring to here.]

For some of these products there already exists similar software, just that they're meant for corporations and are really expensive. Just as an example from something I'm familiar with, for building on Guesstimate there's already Analytica (https://lumina.com/). Now, does it doeverything that Guesstima... (read more)

3
Ozzie Gooen
2y
I agree with (basically) all of this. I've been looking more into enterprise tools for QURI and have occasionally used some. As EA grows, enterprise tools make more sense for us. I guess this seemed to me like a different topic, but I probably should have flagged this somewhere in this post. On Guesstimate in particular, I'm very happy for other groups to use different tools (like Analytica, Causal and probabilistic programming languages. Normally when I talk to people about this, I wind up recommending other options. All that said, I think there are some areas where some smart programming efforts on our behalf could go a long way. I think the space of "monte carlo" tools is still quite small, and I'd like to see other efforts in it (like more startups). One issue with what you mention is that I expect that managing a bunch of corporate licenses will be a pain. It wouldn't be great if some smart people can't see or work with the relevant information because their small team doesn't have a license. So in some situations, it's worth it, but if we can (cheaply) get/use open-source tools and standards, that can also be preferable.

Meta-comment: I noticed while reading this post and some of the comments that I had a strong urge to upvote any comment that was critical of EA and had some substantive content. Introspecting, I think this was partly due to trying to signal-boost critical comments because I don't think we get enough of those, partly because I agreed with some of those critiques, ... but I think mostly because it feels like part of the EA/rationalist tribal identity that self-critiquing should be virtuous. I also found myself being proud of the community that a critical pos... (read more)

I've noted similar patterns, and think that criticisms of EA sometimes get more attention than they deserve. I wrote on related themes here.

(I think this particular post - the OP - makes a good point, however.)

This seems correct and a valid point to keep in mind - but it cuts both ways. It makes sense to reduce your credence when you recognize that expert judgment here is less informed than you originally thought. But by the same token, you should probably reduce your credence in your own forecasts being correct, at least to the extent that they involve inside view arguments like, "deep learning will not scale up all the way because it's missing xyz." The correct response in this case will depend on how much your views depend on inside view arguments about deep ... (read more)

2
Linch
3y
This was my initial reaction, that suspiciousness of existing forecasts can justify very wide error bars but not certainty in >50 year timelines. But then I realized I didn't understand what probability OP gave to <50 years timelines, which is why I asked a clarifying question first. 

Part-time work is an option at my workplace. Less than half-time loses benefits though, which is why I didn't want to drop down to lower than 50%.

  1. I did not have an advisor when I sent the original email, but I did have what amounted to a standing offer from my undergrad ML professor that if I ever wanted to do a PhD he would take me as a grad student. I spent a good amount of time over the past three months deciding whether I should take him up on that or if I should apply elsewhere. I ended up taking him up on the offer.

  2. I did not discuss it with my employer before sending the original email. It did take some work to get it through bureaucratic red tape though (conflict of interest check, etc.).

2
Rory Greig
3y
Very useful to know, thanks for the context!

Does this look close to like what you're looking for? https://www.lesswrong.com/posts/qnA6paRwMky3Q6ktk/modelling-transformative-ai-risks-mtair-project-introduction

If yes, feel free to message me - I'm one of the people running that project.

Also, what software did you use for the map you displayed above?

4
Harrison Durland
3y
What you describe there is probably one of the most similar concepts I've seen thus far, but I think a potentially important difference is that I am particularly interested in a system that allows/emphasizes semantically-richer relationships between concepts and things. From what I saw in that post, it looks like the relationships in the project you describe are largely just "X influences Y" or "X relates to/informs us about  Y", whereas the system I have in mind would allow identifying relationships like "X and Y are inconsistent claims," "Z study had conclusion/finding X," "X supports Y", etc. I used a free/lite/outdated version of Semantica Pro which I was able to get from someone working on it many years ago.

In your 80,000 Hours interview you talked about worldview diversification. You emphasized the distinction between total utilitarianism vs. person-affecting views within the EA community. What about diversification beyond utilitarianism entirely? How would you incorporate other normative ethical views into cause prioritization considerations? (I'm aware that in general this is basically just the question of moral uncertainty, but I'm curious how you and Open Phil view this issue in practice.)

Most people at Open Phil aren't 100% bought into to utilitarianism, but utilitarian thinking has an outsized impact on cause selection and prioritization because under a lot of other ethical perspectives, philanthropy is supererogatory, so those other ethical perspectives are not as "opinionated" about how best to do philanthropy. It seems that the non-utilitarian perspectives we take most seriously usually don't provide explicit cause prioritization input such as "Fund biosecurity rather than farm animal welfare", but rather provide input about what rules... (read more)

True. My main concern here is the lamppost issue (looking under the lamppost because that's where the light is). If the unknown unknowns affect the probability distribution, then personally I'd prefer to incorporate that or at least explicitly acknowledge it. Not a critique - I think you do acknowledge it - but just a comment.

Shouldn't a combination of those two heuristics lead to spreading out the probability but with somewhat more probability mass on the longer-term rather than the shorter term?

5
Ajeya
3y
That's fair, and I do try to think about this sort of thing when choosing e.g. how wide to make my probability distributions and where to center them; I often make them wider than feels reasonable to me. I didn't mean to imply that I explicitly avoid incorporating such outside view considerations,  just that returns to further thinking about them are often lower by their nature (since they're often about unkown-unkowns).
  • What skills/types of people do you think AI forecasting needs?

 

I know you asked Ajeya, but I'm going to add my own unsolicited opinion that we need more people with professional risk analysis backgrounds, and if we're going to do expert judgment elicitations as part of forecasting then we need people with professional elicitation backgrounds. Properly done elicitations are hard. (Relevant background: I led an AI forecasting project for about a year.)

For thinking about AI timelines, how do you go about choosing the best reference classes to use (see e.g., here and here)?

4
Ajeya
3y
I don't think I have a satisfying general answer to this question; in practice, the approaches I pursue first are heavily influenced by which approaches I happen to find some purchase on, since many theoretically appealing reference classes or high-level approaches to the question may be difficult to make progress on for whatever reason.

I know that in the past LessWrong, HPMOR, and similar community-oriented publications have been a significant source of recruitment for areas that MIRI is interested in, such as rationality, EA, awareness of the AI problem, and actual research associates (including yourself, I think). What, if anything, are you planning to do to further support community engagement of this sort? Specifically, as a LW member I'm interested to know if you have any plans to help LW in some way.