M

MichaelDickens

5060 karmaJoined

Bio

I do independent research on EA topics. I write about whatever seems important, tractable, and interesting (to me). Lately, I mainly write about EA investing strategy, but my attention span is too short to pick just one topic.

I have a website: https://mdickens.me/ Most of the content on my website gets cross-posted to the EA Forum.

My favorite things that I've written: https://mdickens.me/favorite-posts/

I used to work as a software developer at Affirm.

Sequences
1

Quantitative Models for Cause Selection

Comments
747

But these should be matched by looking for cases where something good happened because people tried to accumulate power/influence within a system.

I think this is a significant percent of all good things that have ever happened.

I think you are right about this, you've changed my mind (toward greater uncertainty).

My gut feeling is that [...] the biggest difference between good outcomes and bad outcomes is how much work the big AI labs put into alignment during the middle of the intelligence explosion when progress moves fastest.

This seems to depend on a conjunction of several strong assumptions: (1) AI alignment is basically easy; (2) there will be a slow takeoff; (3) the people running AI companies are open to persuasion, and "make AI safety seem cool" is the best kind of persuasion.

But then again I don't think pause protests are going to work, I'm just trying to pick whichever bad plan seems the least bad.

A time cost of 0.0417 $/d for 7.5 s/d and 20 $/h.

Nitpick: I just timed myself taking creatine and it took me 42 seconds.

(My process consists of: take creatine and glass out of cabinet; scoop creatine into glass; pour tap water into glass; drink glass; put creatine and glass back into cabinet.)

Agreed that creatine passes a cost-benefit analysis.

I think this is the best explanation I've seen, it sounds likely to be correct.

That sounds pretty reasonable for why psychotherapy wouldn't be as widespread as it should. It looks to me like most of these reasons wouldn't apply to AMF. Training new psychotherapists takes years and tens of thousands of dollars (at developing-world wages). Getting more malaria nets requires buying more $5 malaria nets, and distributing malaria nets is much easier than distributing psychotherapists. So reasons 1–3 and #6 don't carry over (or at least not to nearly the same extent). #4 doesn't seem relevant to my original question so I think #5 is the only one that carries over—recipients might not know that they should be concerned about malaria.

Why does distributing malaria nets work? Why hasn't everyone bought a bednet already?

  • If it's because they can't afford bednets, why don't more GiveDirectly recipients buy them?
  • Is it because nobody in the local area sells bednets? If so, why doesn't anyone sell them?
  • Is it because people don't think bednets are worth it? If so, why do they use the bednets when given them for free?

I don't know why (I thought it was a good post) but I have some guesses:

  1. Maybe people don't like the framing of o1 as "lying" when it's not clear that it lied. All that's clear is that o1 gave a false justification, which isn't necessarily lying.
  2. You tend to write in an alarmist rhetorical style that I think turns off a lot of people. I think you are very much correct to be alarmed about AI x-risk, but also I don't think it's a good persuasive strategy (for an EA/rationalist audience) to convey this through emotionally charged rhetoric. I didn't think this particular post was alarmist, but you have a history of writing alarmist posts/comments, so maybe people downvoted based on the title.
  • $100 million/year and 600 people = $167,000 per person-year
  • $1M buys 10 person-years = $100,000 per person-year

These numbers are approximately the same. I don't understand how you get that 5/6 of the work comes from volunteering / voluntary underpayment, did I do it wrong?

I plan on donating to PauseAI, but I've put considerable thought into reasons not to donate.

I gave some arguments against slowing AI development (plus why I disagree with them) in this section of my recent post, so I won't repeat those.

  1. There's not that much evidence that protests are effective. There's some evidence, including a few real-world natural experiments and some lab experiments, but this sort of research doesn't have a good replication record.
  2. Research generally suggests that peaceful protests work, while violent protests reduce public support. If violent protests backfire, maybe some types of peaceful protests also backfire. And we don't have enough granularity in the research to identify under what specific conditions peaceful protests work, so maybe PauseAI-style protests will backfire.
  3. Polls suggest there is broad public support for a pause. If there's already public support, doesn't that weaken the case for protesting? Perhaps that means we should be doing something else instead.
  4. Protesting might not be a good use of resources. It might be better to lobby policy-makers.

Yes that's also fair. Conflicts of interest are a serious concern and this might partially explain why big funders generally don't support efforts to pause AI development.

I think it's ok to invest a little bit into public AI companies, but not so much that you'd care if those companies took a hit due to stricter regulations etc.

I think the position I'm arguing for is basically the standard position among AI safety advocates so I haven't really scrutinized it. But basically, (many) animals evolved to experience happiness because it was evolutionarily useful to do so. AIs are not evolved so it seems likely that by default, they would not be capable of experiencing happiness. This could be wrong—it might be that happiness is a byproduct of some sort of information processing, and sufficiently complex reinforcement learning agents necessarily experience happiness (or something like that).

Also: According to the standard story where an unaligned AI has some optimization target and then kills all humans in the interest of pursuing that target (e.g. a paperclip maximizer), it seems unlikely that this AI would experience much happiness (granting that it's capable of happiness) because its own happiness is not the optimization target.

(Note: I realize I am ignoring some parts of your comment, I'm intentionally only responding to the central point so my response doesn't get too frayed.)

Load more