All of Peter's Comments + Replies

  1. Interesting. Are there any examples of what we might consider a relatively small policy changes that received huge amounts of coverage? Like for something people normally wouldn't care about. Maybe these would be informative to look at compared to more hot button issues like abortion that tend to get a lot of coverage. I'm also curious if any big issues somehow got less attention than expected and how this looks for pass/fail margins compared to other states where they got more attention. There are probably some ways to estimate this that are better than o
... (read more)

I'd be curious to hear about potential plans to address any of these, especially talent development and developing the pipeline of AI safety and governance. 

2
michel
5mo
Any plans to address these would come from the individuals or orgs working in this space. (This event wasn't a collective decision-making body, and wasn't aimed at creating a cross-org plan to address these—it was more about helping individuals refine their own plans).  Re the talent development pipeline for AI safety and governance, some relevant orgs/programs I'm aware of off the top of my head include: * Arkrose * SERI MATS * Constellation * Fellowships like AI Futures * Blue Dot Impact * AI safety uni groups like MAIA and WAISI * ... and other programs mentioned on 80K job board and EA Opportunity Board

Very interesting. 
1. Did you notice an effect of how large/ambitious the ballot initiative was? I remember previous research suggesting consecutive piecemeal initiatives were more successful at creating larger change than singular large ballot initiatives. 

2. Do you know how much the results vary by state?

3. How different do ballot initiatives need to be for the huge first advocacy effect to take place? Does this work as long as the policies are not identical or is it more of a cause specific function or something in between? Does it have a smooth gradient or is it discontinuous after some tipping point?

2
zdgroff
5mo
1. I look at some things you might find relevant here. I try to measure the scale of the impact of a referendum. I do this two ways. I have just a subjective judgment on a five-point scale, and then I also look at predictions of the referendum's fiscal impact from the secretary of state. Neither one is predictive. I also look at how many people would be directly affected by a referendum and how much news coverage there was before the election cycle. These predict less persistence. 2. This is something I plan to do more, but they can't vary that much because when I look at variables that vary across states (e.g., requirements to get on the ballot), I don't see much of a difference. 3. I'm not totally sure what your question is, but I think you might be interpreting my results as saying that close referendums are especially persistent. I'm only focusing on close referendums because it's a natural experiment—I'm not saying there's something special about them otherwise. I'm just estimating the effect of passing a marginal referendum on whether the policy is in place later on. I can try to think about whether this holds for things that are not close by looking at states with supermajority requirements or by looking at legislation, and it looks like things are similar when they're not as close.

This is an inspiring amount of research. I really appreciate it and am enjoying reading it. 

That's a good point. Although 1) if people leave a company to go to one that prioritizes AI safety, then this means there are fewer workers at all the other companies who feel as strongly. So a union is less likely to improve safety there. 2) It's common for workers to take action to improve safety conditions for them, and much less common for them to take action on issues that don't directly affect their work, such as air pollution or carbon pollution, and 3) if safety inclined people become tagged as wanting to just generally slow down the company, then hiring teams will likely start filtering out many of the most safety minded people. 

I've thought about this before and talked to a couple people in labs about it. I'm pretty uncertain if it would actually be positive. It seems possible that most ML researchers and engineers might want AI development to go as quickly or more than leadership if they're excited about working on cutting edge technologies or changing the world or for equity reasons. I remember some articles about how people left Google for companies like OpenAI because they thought Google was too slow, cautious, and lost its "move fast and break things" ethos. 

1
dEAsign
7mo
As you have said there are examples of individuals have left firms because they feel their company is too cautious. Conversely there are individuals who have left for companies that priorities AI safety. If we zoom out and take the outside view, it is common for those individuals who form a union to take action to slow down or stop their work or take action to improve safety. I do not know an example of a union that has instead prioritised acceleration.

Really appreciate this post. Recently I've felt less certain about whether slowing down AI is feasible or helpful in the near future. 

I think how productive current alignment and related research is at the moment is a key crux for me. If it's actually quite valuable at the moment, maybe having more time would seem better. 

It does seem easier to centralize now when there are fewer labs and entrenched ways of doing things, though it's possible that exponentially rising costs could lead to centralization through market dynamics anyway. Though maybe that would be short lived and some breakthrough after would change the cost of training dramatically. 

Yes, it seems difficult to pin those down. Looking forward to the deeper report!

I really want to see more discussion about this. There's serious effort put in. I've often felt that nuclear is perhaps overlooked/underemphasized even within EA. 

2
Joel Tan
1y
The expected disvalue is really high, especially compared to other longtermist risks, where the per annum probabilities of bad stuff happening is fundamentally low! The worry, I think, is concentrated on how tractable any intervention is, in a context where it's hard to know the chances of success before the fact, and about as hard to do attribution after.

Actually, they are the same type of error. EA prides itself on using evidence and reason rather than taking the assessments of others at face value. So the idea that others did not sufficiently rely on experts who could obtain better evidence and reasoning to vet FTX is less compelling to me as an after-the-fact explanation to justify EA as a whole not doing so. I think probably just no one really thought much about the possibility and looking for this kind of social proof helps us feel less bad. 

Yeah, I do sometimes wonder if perhaps there's a reason we find it difficult to resolve this kind of inquiry. 

Yes, I think they're generally pretty wary of saying much exactly since it's sort of beyond conceptual comprehension. Something probably beyond our ideas of existence and nonexistence. 

Glad to hear that! You're welcome :)

On Flynn Campaign: I don't know if it's "a catastrophe" but I think it is maybe an example of overconfidence and naivete. As someone who has worked on campaigns and follows politics, I thought the campaign had a pretty low chance of success because of the fundamentals (and asked about it at the time) and that other races would have been better to donate to (either state house races to build the bench or congressional candidates with better odds like Maxwell Frost, a local activist who ran for the open seat previously held by Val Demings, listed pandemic pr... (read more)

5
RyanCarey
1y
In the first example, you complain that EA neglected typical experts and "EA would have benefited from relying on more outside experts" but in the second example, you say that EA "prides itself on NOT just doing what everyone else does but using reason and evidence to be more effective", so should have realised the possible failure of FTX. These complaints seem exactly opposite to one another, so any actual errors made must be more subtle.

I think the main obstacle is tractability: there doesn't seem to be any known methodology that could be applied to resolve this question in a definitive way. And it's not clear how we could even attempt to find such a method. Whereas projects related to areas such as preventing pandemics and making sure AI isn't misused or poorly designed seem 1) incredibly important, 2) tractable - it looks like we're making some progress and have and can find directions to make further progress (better PPE, pathogen screening, new vaccines, interpretability, agent founda... (read more)

2
Garrett Ehinger
1y
I agree with @MikeJohnson on thought experiments falling within a deist frame (such as Nick Bostrom's Simulation Hypothesis), however I'd hardly say these make TI tractable. I'd rather say that research into quantum consciousness or string theory etc. have very strong scientific bases and I personally think they have set good precedents for concluding TI. I.e., they make a good case for just how tractable TI can be. A good book that sums this up pretty well is Jeffrey M. Schwartz M.D.'s "The Mind and the Brain". He goes into the implications of quantum consciousness and the potential for there to be Creator's that we could possibly be influenced by via String Theory related physics, and that this could be tested for. I think people would be surprised by just how tractable this could be, but honestly it's contingent on the nature of a Creator if that Creator does exist. Like I said in the last clause of my post, if the Creator's don't want to be found or are impossible to observe, then we are wasting our time no matter how theoretically tractable TI might be, so ultimately I have to say I sort of agree with your point, Peter!   As for your point on impermanence, I'm pretty sure every religion believes that everything continues forever; although some do get nuanced regarding whether or not that "forever" is divided up into infinite separate lives like the aforementioned Buddhists, but even they believe that once you've obtained complete enlightenment and have shed your Karma you exit the cycle of 轮回 (lun hui) and enter an eternal state of peace. The only group of people I can think of who don't believe something along the lines of an eternal afterlife in a heaven or hell world are die-hard heat-death atheists, which is a pretty small subset of the atheist population if I'm correct. Ultimately, its still a part of TI that deserves answering I think. As for your last point, I definitely see the merit of your point there! Thanks a bunch for sharing that! It's an awesom
8
MikeJohnson
1y
Speaking broadly, I think people underestimate the tractability of this class of work, since we’re already doing this sort of inquiry under different labels. E.g., 1. Nick Bostrom coined, and Roman Yampolskiy has followed up on, the Simulation Hypothesis, which is ultimately a Deist frame; 2. I and others have written various inquiries about the neuroscience of Buddhist states (“neuroscience of enlightenment” type work); 3. Robin Hanson has coined and offered various arguments around the Great Filter. In large part, I don’t think these have been supported as longtermist projects, but it seems likely to me that there‘s value in pulling these strings, and each is at least directly adjacent to theological inquiry.

Thank you - I had forgotten about that post and it was really helpful. 

I've definitely seen well-meaning people mess up interactions without realizing it in my area (non-EA related). This seems like a really important point and your experience seems very relevant given all the recent talk about boards and governance. Would love to hear more of your thoughts either here or privately. 

Seems interesting, I'll def check it out sometime 

Jokes  aside, this is a cool idea. I wonder if reading it yourself and varying the footage, or even adapting the concepts into something  would help it be more attractive to watch. Though of course these would all increase the time investment cost. I can't say it's my jam but I'd be curious to see how these do on TikTok though since they seem to be a sort of prevalent genre/content style. 

Yeah I think college students will often think "Fellowship" is religious because that's likely the only context they have seen the word used in, even though it's often used for all kinds of non-religious opportunities. 

I'm not sure how important this is - I soon realized lots of fellowships at my school were not religious and that it had a broader meaning. 

I guess people could try different things out and see how they work. Maybe something simple like EA reading group. Or focus on a topic name: people would probably be less likely to confuse something like "public health/pandemic prevention/AI Ethics fellowship" as something religious. 

I have thought a few times that maybe a safer route to AGI would be to learn as much as we can about the most moral and trustworthy humans we can find and try to build on that foundation/architecture. I'm not sure how that would work with existing convenient methods of machine learning. 

Yeah there are a lot of "fairweather friends" in politics who won't feel inclined to return any favors when it matters most. The opposite of that is having a committed constituency that votes enough in elections to not be worth upsetting - aka a base of people power. These take serious effort to create and not all groups are distributed geographically the same way so some have more/easier influence than others. One reason the NRA is so powerful and not abandoned despite negative media coverage is that they have tight relationships with Republican politicia... (read more)

I think politics can seem very opaque, incomprehensible, and lacking clear positive payoffs but after volunteering, studying, and working on campaigns for a few years, I think it's more simple but difficult. 

  1.  I think politics is an area where there are a lot of entrenched ways of doing things as well as a lot of pitfalls that often require experience to navigate well. And even then, the chance of failure is still high. A moment's slip up or bad assumptions or a random event can undermine months or years of work. This doesn't happen as often in ot
... (read more)
3
trevor1
1y
In addition to blowback, there's also the risk of getting entangled into specific political factions and then tanking with them, or even having blame deliberately passed to EA associates because they're vulnerable and/or make for a convenient villain.  These people are essentially armies of lawyers (sometimes literally) and they have developed substantial cost effectiveness at selecting targets and spinning narratives, and they often randomly run into hurdles that incentivize them to think up all sorts of ways to cut their losses.

If you click on your name in the top right corner, then click edit profile, you can scroll down and delete tags under "my activity" by clicking the x on the right side of each block.

2
Charles He
1y
Yes! Thank you!

What things would make people less worried about AI safety if they happened? What developments in the next 0-5 years should make people more worried if they happen?

4
plex
1y
on the good side 1. We hit some hard physical limit on computation which dramatically slows the relevant variants of Moore's law. 2. A major world power wakes up and started seriously focusing on AI safety as a top priority. 3. We build much better infrastructure for scaling the research field (e.g. AI tools to automatically connect people with relevant research, and help accelerate learning with contextual knowledge of what a person has read and is interested in, likely in the form of a unified feed (funding offers welcome!)). Apart's AI Safety Ideas also falls into the category of key infrastructure. 4. Direct progress on alignment, some research paradigm emerges which seems likely to result in a real solution. 5. Dramatic slow in the rate of capabilities breakthroughs, discovering some crucial category of task which needs fundamental breakthroughs which are not quick to be achieved. 6. More SBFs entering the funding ecosystem. and on the bad side 1. Lots of capabilities breakthroughs. 2. Stagnation or unhealthy dynamics in the alignment research space (e.g. vultures, loss of collective sensemaking as we scale, self-promotion becoming the winning strategy). 3. US-China race dynamics, especially both countries explicitly pushing for AGI without really good safety considerations. 4. Funding crunch.

What are good ways to test your fit for technical AI Alignment research? And which ways are best if you have no technical background?

4
Iyngkarran Kumar
1y
This is a really comprehensive post on pursuing a career in technical AI safety, including how to test fit and skill up 

Well squad-esque seems like an odd litmus test since there are many other progressive members of congress than them but POF did support Maxwell Frost who won. 

Well to be fair I didn't say it was impossible, just that the outcome probably had more to do with the fundamentals of the race. It may have had a negative effect yes, but plenty of candidates win in races despite being supported by all kinds of PACs and having negative press about it. 

Having more connections within the state for support and donations and highlighting those would have helped blunt negative attacks about PAC funding, for example. 

I like the idea of Protect Our Future being more transparent about how and why they make endorsements. Giving a specific list of ways they evaluate candidates would be helpful for people to understand their actions. I also worry a little bit that this would make it easy to game their endorsement process or encourage political stunts that are more about drawing attention than doing something useful. But I'm not sure how big of a worry this should be.

Not sure but I think the Flynn campaign result was more likely an outcome of the fundamentals of the race: a popular, progressive, woman of color with local party support who already represented part of the district as a state rep and helped draw the new congressional district was way more likely to win over someone who hadn't lived there in years and had never run a political campaign before. 

1
ChristianKleineidam
1y
Andrea Salinas got 36% of the vote while Carrick Flynn got 18%. I think it's pretty clear, that Flynn would have gotten more votes if he wouldn't have been perceived by the press as being funded by ill-intentioned corporate money.  Whether that would have been enough to get double the amount of votes is unclear but I don't think the available data suggest that this isn't in the realm of what would have been possible. 

Hi everyone, I'm a psychology graduate interested in learning about ways for dealing with infohazards especially in online journal publications. As well as what skills and projects are important in Bioethics. 

In terms of goal directedness, I think a lot of the danger hinges on whether and what kinds of internal models of the world will emerge in different systems, and not knowing what those will look like. Many capabilities people didn't necessarily expect or foresee suddenly emerged after more training - for example the jumps in abilities from GPT to GPT-2 and GPT-3. A similar jump to the emergence of internal models of the world may happen at another threshold. 

I think I would feel better if we had some way of concretely and robustly specifying "goal dir... (read more)

Oh thanks! I think I will once I get them a bit more organized. 

I have some ideas and I'm starting to flesh them out so yeah!

2
Devin Kalish
2y
Nice! I'm pretty busy right now so I can't guarantee efficiency, but feel free to send me any outlines/drafts if you would like feedback.

Yeah I can see what you mean - they could have taken a less flashy and more straightforward approach. It would be interesting to think about what else they could have made or done with what they made instead that might have been better. 

Yes - I didn't think about looking at how humane laws for people correlate with animals though. That's really interesting. 

Do you write fiction at all?

 

1
Devin Kalish
2y
No not really, I haven’t written any fiction in a long while, I mostly just blog now. And you?

Thanks for sharing. Yeah, I can see what you mean - Senku can be a bit annoying. I think that also makes him a more realistic character too, though maybe they overdo it a bit at times. I found Gen's syllable switching (like saying the second half first) on certain words really grating and it seemed to come out of nowhere. I didn't remember him doing that at the beginning. I think it would be super cool to try and make more stories like Dr. Stone where it's entertainment but also learning real stuff is a tool the characters use to progress the plot. 

Oh... (read more)

1
Devin Kalish
2y
If I recall the cellphone mostly had a substantial role in the war when Gen pretended to be Lillian, which I don't think was part of that plan from the beginning, and doesn't seem like it was a priori all that likely to work. This might be a nit pick, but it's just still hard for me to see a realistic strategy in which spending all that time on a cellphone was worth skipping out on any other technology they could have used that time to make, even for a peaceful resolution. I kind of see that part as having just been to show off how you could make such an advanced technology with such primitive resources, which is cool but sticks out without some better in-canon justification. Also, is this roughly what you were thinking of with the overlap between animal rights and human rights?

Oh yeah that makes sense, I agree. Yeah, FMAB is often a good "first anime" to recommend since it does lots of things pretty well.

I'm really curious, how would you improve Dr. Stone? I think it could be improved but I'm not overflowing with ideas on how to do it at the moment. 

Oh, I forgot about him being vegetarian. I think the reason the AI angle is more popular is because of how much more similar he seems to be to humans than animals. There are so many qualities people think of as being human capabilities/behavior that he does even if not all of th... (read more)

2
Devin Kalish
2y
If you think it’s alright, I wouldn’t mind continuing the conversation here, though feel free to jump in and tell me if it bothers you Luca. On the point of Dr Stone’s flaws, I don’t think it is incredibly flawed overall, there are other animes I like, including much more popular ones like Attack on Titan and Demon Slayer, which I think are significantly more flawed. I mostly just think that FMAB is an outlier in its polish, and is well executed beyond this. I think many of the things I think of as Dr Stone’s flaws are fairly minor in the grand scheme, for instance Senku can be even more obnoxious about his intelligence than HPMOR Harry, to the point of coming off as pretty mean to his best friend sometimes, though this improves a bit over time. There are also pretty overused catchphrases like “10 billion”, “this is bad”, and Gen’s pig latin that get annoying. Maybe more substantially there are moments of distractingly inconsistent rigor/rationality, for instance the stone wars arc when they spend months on making a cell phone only to make a passable tank as an afterthought. Again though, I like Dr Stone, and although it comes off to me as rougher around the edges, I don’t think it’s egregiously flawed. On the point of Frankenstein, I think that the connection to animal rights and AI is pretty similar overall, the creature is both literally an artificial intelligence and literally a non-human animal, but bears little resemblance to how either is portrayed elsewhere in fiction and serious discussions. Cynically, the reason the animal rights connection didn’t take off might have something to do with the animal rights being less popular. Less cynically, writing about AI was more unique at the time than writing about non-humans, so it left a deeper impression. Either way, I think both provide worthwhile readings, even if I prefer the animal rights one.

"They genuinely weren't surprised by anything that happened. They didn't necessarily predict everything perfectly, but everything that happened matched their model well enough. Their deep insight into ML progress enables them to clearly explain why AGI isn't coming soon, and they can provide rough predictions about the shape of progress over the coming years."

Would definitely like to hear from people like this and see them make lots of predictions about AI progress in the short term (months and years). 

Seems like a very promising profile for identifying those with a valuable counter perspective/feedback for improving ideas. 

FMAB is pretty widely liked. It used to be my favorite, but nowadays I think I put more emphasis on things that change the way I think. It has been a long time since I watched it though so I might change my mind if I rewatched. 

Frankenstein makes me think about AI as well since it's all about creating something with greater capabilities than a human. 

I've been meaning to read The Dispossessed. Will have to check out those other ones. 

1
Devin Kalish
2y
I more or less agree with that on FMAB, I think it’s a really well done show, but didn’t significantly change the way I think. As a rule most of the anime I’ve seen has some sort of serious flaws in execution or content, much of the appeal I get from anime is shows that are somewhat more unique in their strengths but also have more weaknesses, like Dr Stone. FMAB is unusually well executed which makes it sort of easier to point to as the overall best, it is just a basically strong story, even though I probably enjoy shows like Dr Stone better as a rule. I think Frankenstein does connect to AI writing, and that may be the genre it’s had the most overall influence on, but personally I think the connection between this and current AI issues is fairly weak and may have taught the genre some of the wrong lessons. I connect it more directly to animal rights actually. The philosophical connection is pretty clear, because most contemporary animal rights arguments start with the rejection of the moral importance of mere species membership. This is played out in the content to some extent as well, for instance the creature is canonically a vegetarian or vegan, “My food is not that of man; I do not destroy the lamb and the kid, to glut my appetite; acorns and berries afford me sufficient nourishment.” and even plausibly a utilitarian, “I felt the greatest ardour for virtue rise within me, and abhorrence for vice, as far as I understood the signification of those terms, relative as they were, as I applied them, to pleasure and pain alone.” (sorry I don’t have page numbers, I’m working from Project Gutenberg), and there are various other points in this and other pieces of Shelley’s work that hint at care for animals. There is also circumstantial evidence for this concern, her husband Percy Shelley, along with her deceased mother Mary Wollstonecraft whom she idolized, were both concerned with animal rights. She even had some indirect connections to the philosophical radica
2
Devin Kalish
2y
Both are pretty difficult. With anime in particular there are plenty I like isolated things about, but my favorite all things considered is probably just a really well done one overall like Fullmetal Alchemist: Brotherhood or Deathnote. As for science fiction broadly, for a long time my favorite novel was Daniel Keyes' Flowers for Algernon, though I haven't read it in a while. The Dispossessed by Le Guin comes to mind as a very well done one too, but I'm leaning more towards Frankenstein lately, which is more personal because I wrote my capstone on it, and its connections to my current ethical interests. Short fiction is probably easier, my favorite short story is probably Thirteen Ways of Looking at Space/Time by Catherynne M Valente, which is sort of science fiction.

Oh your post is probably what got it on my radar then -  thanks!

1
Devin Kalish
2y
Oh wow, I'm glad to hear it!
Answer by PeterOct 11, 20221
0
0

I liked Dr. Stone and Madoka Magica. The first one is pretty good at being entertainment that also occasionally happens to teach you about practical science concepts and the impacts of new technology, and the latter thinking about what we value and why. 

Overlord makes me think about AI a lot. Having a bunch of undead soldiers that can do stuff like farm 24/7 without getting tired or needing to eat is kind of like having robots that can automate work. It also completely unbalances the world economy. And having a bunch of powerful servants who were crea... (read more)

1
Devin Kalish
2y
I just answered a similar question on another post before realizing the post was over a year old and pretty quiet. But yeah, the main takeaway is that I second Dr. Stone, I've written about its connection to EA a couple times at this point, and while it's not my favorite piece of science fiction, or even probably my favorite anime, I think it is a good fun show that's unusually relevant to EA.

This is really useful and makes sense - thanks for sharing your findings! 

In my experience talking about an existing example of a problem like recommendation systems prioritizing unintended data points + example of meaningful AI capability usually gets people interested. Those two combined would probably be bad if we're not careful. Jumping to the strongest/worst scenarios usually makes people recoil because it's bad and unexpected and doesn't make sense why you're jumping to such an extreme outcome.

Do you have any examples of resources you were unaware of before? That could be useful to include as a section both for the actual resources and thinking about how to find such sources in the future. 

2
mariushobbhahn
2y
Reflects my experience! The resources I was unaware of were usually highly specific technical papers (e.g. on some aspect of interpretability), so nothing helpful for a general audience.
Answer by PeterOct 08, 20223
1
0

I doubt it's the most important, and maybe it's not a good idea, but an option for users to hide post and comment karma from their own point of view could be interesting. 

I think outreach to Buddhists is an interesting idea! I think many Buddhists would probably agree that the point of practice is to become a better person and thus have more impact on the world. And there are probably some who don't necessarily think about it the same way. 

Buddhists and EAs could probably learn some things from each other yes. This is generally a good attitude to have towards others - curiosity and openness to experimentation. 

I do think it's probably important for some people to dedicate themselves fulltime to practice and teaching as a profession, to keep the teachings and practices alive in a high-fidelity way. 

2
maximizealtruism
2y
Would love to speak more about it on EARadio sometime :)

I think building skills to become better at research and building skills to become better at things like dealing with stress or interacting with others are both important to having a greater positive impact on the world, so my point wasn't exactly about deontology vs consequentialism. 

And I'd guess EA probably has more concrete consensus on the former than the latter.  

This sounds interesting. One worry I have would be preventing any kind of exploitation of recipients in exchange for support. 

Oh I can see what you mean. This post probably already took so many hours to make. Thanks for your response. 

I think another part is that there definitely seems to be some variety of opinion on how targeted EA should be vs broad to maximize impact. I think I lean more towards broader, since 100 people doing their best is probably better than 10, even if the 10 are in 5x  more high impact roles/situations. I have some other half baked thoughts probably but I should take some time to think them through.  

Yeah, that's a good point about giving. Giving itself can be a transformative action. I guess if I were to put it another way, EA shares many religions' emphasis on works/charity but not necessarily experiences and practices of 'faith'/transcendence. 

In terms of the Mormon stuff, I think importing cultural habits is maybe different but adjacent to what I mean. Although maybe some aren't that separate. It seems like abstaining from alcohol exists in a lot of different religions so maybe that particular behavior was found to be helpful in some way to pe... (read more)

Load more