All of mic's Comments + Replies

See Ask MIT Climate: Why do some people call climate change an “existential threat”?

That more literal-minded reading of the phrase “existential threat” may not be the best reflection of the risks of climate change, however. “Even under our most dire predictions, human society is still around,” says Adam Schlosser, the Deputy Director of the MIT Joint Program on the Science and Policy of Global Change and a climate scientist who studies future climate change and its impact on human societies. “I do not personally view this as an extinction issue. But there

... (read more)

In 2018, the Center for Humane Technology's "Time Well Spent" campaign probably contributed to Apple's Screen Time and Google's Digital Wellbeing features. These features seem deliberately hampered because of how easy it is to disable the reminder for you to get off your phone. I wonder if this problem is hard to make real traction on because tech companies, especially ones that make revenue from advertisements, are actively motivated against reducing screen time.

Great piece overall! I'm hoping AI risk assessment and management processes can be improved.

Anthropic found that Claude 3 didn't trigger AI Safety Level 3 for CBRN, but gave it a 30% chance of doing so in three months

30% chance of crossing the Yellow Line threshold (which requires building harder evals), not ASL-3 threshold

1
Luca Righetti 🔸
Thanks for the spot! Have fixed
mic
21
3
4
1

The plant-based foods industry should make low-phytoestrogen soy products.

Soy is an excellent plant-based protein. It's also a source of the phytoestrogen isoflavone, which men online are concerned has feminizing properties (cf. soy boy). I think the effect of isoflavones is low for moderate consumption (e.g., one 3.5 oz block of tofu per day), but could be significant if the average American were to replace the majority of their meat consumption with soy-based products.

Fortunately, isoflavones in soy don't have to be an issue. Low-isoflavone products are ... (read more)

3
Milan Weibel🔹
I think it would be really hard (maybe even practically impossible) to market isoflavone-reduced products without hurting demand for non-isoflavone-reduced products as a side effect.  If the plant-based food industry started producing and marketing isoflavone-reduced soy products, I am quite confident that it would counterfactually lower total demand for soy products in the short term, and I am very uncertain about the sign of impact over the long term.
8
akash 🔸
Could you elaborate how you conclude that the effects of soy isoflavones could be significant if consumption were higher? I read this summary article from the Linus Pauling institute a while ago and concluded, "okay, isoflavones don't seem like an issue at all, and in some cases might have health benefits" (and this matches my experience so far).[1] The relevant section from the article: Unless there is some new piece of information that fairly moderately/strongly suggests that isoflavones do have feminizing effects, this seems like a non-issue.  1. ^ A personal anecdote, not that it bears much weight, I have been consuming >15 ounces of tofu and >250 ml of soy milk nearly every day for the last four years, and I have noticed how "feminine" or "masculine" my body looks is almost entirely dependent on how much weight I lift in a week and my nutritional intake, rather than my soy intake.

Animal welfare does more to push the frontiers of moral circle expansion

mic
20
2
0
1

Harris was the one personally behind the voluntary AI safety commitments of July 2023. Here's a press release from the White House:

The Vice President’s trip to the United Kingdom builds on her long record of leadership to confront the challenges and seize the opportunities of advanced technology. In May, she convened the CEOs of companies at the forefront of AI innovation, resulting in voluntary commitments from 15 leading AI companies to help move toward safe, secure, and transparent development of AI technology. In July, the Vice President convened

... (read more)

I'm surprised the video doesn't mention cooperative AI and avoiding conflict among transformative AI systems, as this is (apparently) a priority of the Center on Long-Term Risk, one of the main s-risk organizations. See Cooperation, Conflict, and Transformative Artificial Intelligence: A Research Agenda for more details.

mic
9
2
0
1

I wouldn't consider factory farming to be an instance of astronomical suffering, as bad as the practice is, since I don't think the suffering from one century of factory farming exceeds hundreds of millions of years of wild animal suffering. However, perhaps it could be an s-risk if factory farming somehow continues for a billion years. For reference, here is definition of s-risk from a talk by CLR in 2017:

“S-risk – One where an adverse outcome would bring about severe suffering on a cosmic scale, vastly exceeding all suffering that has existed on Earth so far.”

Great to see this! One quick piece of feedback: It takes a while to see a response from the chatbot. Are you planning on streaming text responses in the future?

3
Animal_Ethics
Yes, and we are also planning to switch to Claude Haiku (a faster model for generating responses).

Thanks for your comment, Jackson! I've removed my post since it seems that it was too confusing. One message that I meant to convey is that the imaginary nuclear company essentially does not have any safety commitments currently in effect ("we aren't sure yet how to operate our plant safely") and is willing to accept any number of deaths less than <10,000 people, despite adopting this "responsible nuclear policy."

I think another promising intervention would be to persuade God to be a conditional annihilationist or support universal reconciliation with Christ. Abraham successfully negotiated conditions with God regarding the destruction of Sodom and Gomorrah with just a few sentences. Imagine what we could do with rigorous and prayerful BOTEC analyses! Even if there is a small chance of this succeeding, the impact could be incredible in expectation.

Are there any readings about how a long reflection could be realistically and concretely achieved?

mic
11
1
0

Great post! I've written a paper along similar lines for the SERI Conference in April 2023 here, titled "AI Alignment Is Not Enough to Make the Future Go Well." Here is the abstract:

AI alignment is commonly explained as aligning advanced AI systems with human values. Especially when combined with the idea that AI systems aim to optimize their world based on their goals, this has led to the belief that solving the problem of AI alignment will pave the way for an excellent future. However, this common definition of AI alignment is somewhat idealistic and mis

... (read more)
mic
5
3
0
1

Congrats on launching GWWC Local Groups! Community building infrastructure can be hard to set up, so I appreciate the work here.

It would be bad to create significant public pressure for a pause through advocacy, because this would cause relevant actors (particularly AGI labs) to spend their effort on looking good to the public, rather than doing what is actually good.

I think I can reasonably model the safety teams at AGI labs as genuinely trying to do good. But I don't know that the AGI labs as organizations are best modeled as trying to do good, rather than optimizing for objectives like outperforming competitors, attracting investment, and advancing exciting capabilities – subjec... (read more)

I don't know that the AGI labs as organizations are best modeled as trying to do good, rather than optimizing for objectives like outperforming competitors, attracting investment, and advancing exciting capabilities – subject to some safety-related concerns from leadership.

I will go further -- it's definitely the latter one for at least Google DeepMind and OpenAI; Anthropic is arguable. I still think that's a much better situation than having public pressure when the ask is very nuanced (as it would be for alignment research).

For example, I'm currently gla... (read more)

I think GiveWell shouldn’t be modeled as wanting to recommend organizations that save as many current lives as possible. I think a more accurate way to model them is “GiveWell recommends organizations that are [within the Overton Window]/[have very sound data to back impact estimates] that save as many current lives as possible.”

This is correct if you look at GiveWell's criteria for evaluating donation opportunities. GiveWell’s highly publicized claim “We search for the charities that save or improve lives the most per dollar” is somewhat misleading g... (read more)

Upvoted. This is what longtermism is already doing (relying heavily on non-quantitative, non-objective evidence) and the approach can make sense for more standard local causes as well.

What do you think are the main reasons behind wanting to deploy your own model instead of training an API? Some reasons I can think of:

For anyone interested, the Center for AI Safety is offering up to $500,000 in prizes for benchmark ideas: SafeBench (mlsafety.org)

Where do you draw the line between AI startups that do vs don't contribute excessively to capabilities externalities and existential risk? I think you're right that your particular startup wouldn't have a significant effect of accelerating timelines. But if we're thinking AI startups in general, this could be another OpenAI or Adept, which probably have more of an effect on timelines.

I could imagine that even if one's startup doesn't working on scaling and making models generally smarter, a relatively small amount of applications work to make them more use... (read more)

3
Joshc
If your claim is that 'applying AI models for economically valuable tasks seems dangerous, i.e. the AIs themselves could be dangerous' then I agree. A scrappy applications company might be more likely to end the world than OpenAI/DeepMind... it seems like it would be good, then, if more of these companies were run by safety conscious people. A separate claim is the one about capabilities externalities. I basically agree that AI startups will have capabilities externalities, even if I don't expect them to be very large. The question, then, is how much expected money we would be trading for expected time and what is the relative value between these two currencies.

Just so I understand, are all four of these quotes arguing against preference utilitarianism?

1
Kaspar Brandner
Yes exactly.

I'm curious whether the reason why EA may be perceived as a cult while, e.g., environmentalist and social justice activism are not, is primarily that the concerns of EA are much less mainstream.

I appreciate the suggestions on how to make EA less cultish, and I think they are valuable to implement, but I don't think they would have a significant effect on public perception of whether EA is a cult.

I agree, that seems concerning. Ultimately, since the AI developers are designing the AIs, I would guess that they would try to align the AI to be helpful to the users/consumers or to the concerns of the company/government, if they succeed at aligning the AI at all. As for your suggestions "Alignment with whoever bought the AI? Whoever users it most often? Whoever might be most positively or negatively affected by its behavior? Whoever the AI's company's legal team says would impose the highest litigation risk?" – these all seem plausible to me.

On the sepa... (read more)

But I sometimes have a fear in the back of my mind that some of the attendees who are intrigued by these ideas are later going to look up effective altruism, get the impression that the movement’s focus is just about existential risks these days, and feel duped.  Since EA pitches don’t usually start with longtermist ideas, it can feel like a bait and switch.

To avoid the feeling of a bait and switch, I think one solution is to introduce existential risk in the initial pitch. For example, when introducing my student group Effective Altruism at Georgia T... (read more)

1
JoshuaTheLion
I think this is a great opening pitch! Not every personality type is well-equipped to work in AI alignment, so I strongly feel the pitch should be about finding the field that best suits you and where you specifically can find your greatest impact, regardless of whether that ends up being in a longtermist career or global health/poverty or earning to give. As to what charity to give to, whether it would be better to donate to a GiveWell charity or better to donate to AI alignment research, I personally am not sure, I lean more toward the GiveWell charities but I'm fairly new to the EA community and still forming my opinions....

I think AI alignment isn't really about designing AI to maximize for the preference satisfaction of a certain set of humans. I think an aligned AI would look more like an AI which:

  • is not trying to cause an existential catastrophe or take control of humanity
  • has had undesirable behavior trained out or adversarially filtered
  • learns from human feedback about what behavior is more or less preferable
    • In this case, we would hope the AI would be aligned to the people who are allowed to provide feedback
  • has goals which are corrigible
  • is honest, non-deceptive, and non-power-seeking
4
Geoffrey Miller
Hi mic, I understand that's how 'alignment' is normally defined in AI safety research.  But it seems like such a narrow notion of alignment that it glosses over almost all of the really hard problems in real AI safety -- which concern the very real conflicts between the humans who will be using AI. For example, if the AI is aligned 'to the people who are allowed to provide feedback' (eg the feedback to a CIRL system), that raises the question of who is actually going to be allowed to provide feedback. For most real-world applications, deciding that issue is tantamount to deciding which humans will be in control of that real-world domain -- and it may leave the AI looking very 'unaligned' to all the other humans involved.

Thanks for writing this! There's been a lot of interest of EA community building, but I think the one of the most valuable parts of EA community building is basically just recruiting – e.g., notifying interested people about relevant opportunities and inspiring people to apply for impactful opportunities. A lot of potential talent isn't looped in with a local EA group or the EA community at all, however, so I think more professional recruiting could help a lot with solving organizational bottlenecks.

I was excited to read this post! At EA at Georgia Tech, some of our members are studying industrial engineering or operations research. Should we encourage them to reach out to you if they're interested in getting involved with operations research for top causes?

2
wesg
Yes, definitely!

What are some common answers you hear for Question #4: "What are the qualities you look for in promising AI safety researchers? (beyond general intelligence)"

mic
35
8
0

Technical note: I think we need to be careful to note the difference in meaning between extinction and existential catastrophe. When Joseph Carlsmith talks about existential catastrophe, he doesn't necessarily mean all humans dying;  in this report, he's mainly concerned about the disempowerment of humanity. Following Toby Ord in The Precipice, Carlsmith defines an existential catastrophe as "an event that drastically reduces the value of the trajectories along which human civilization could realistically develop". It's not straightforward to translat... (read more)

5
quinn
I have suggested we stop conflating positive and negative longtermism. I found, for instance, the Precipice hard to read because of the way he flipped back and forth between the two.
6
Algo_Law
I face enormous challenges convincing people of this.  Many people don't see, for example, widespread AI-empowered human rights infringements as an 'existential catastrophe' because it doesn't directly kill people, and as a result it falls between the cracks of AI safety definitions - despite being a far more plausable threat than AGI considering it's already happening. Severe curtailments to humanity's potential still firmly count as an existential risk in my opinion.
mic
11
0
0

Here's my proposal for a contest description. Contest problems #1 and 2 are inspired by Richard Ngo's Alignment research exercises.

AI alignment is the problem of ensuring that advanced AI systems take actions which are aligned with human values. As AI systems become more capable and approach or exceed human-level intelligence, it becomes harder to ensure that they remain within human control instead of posing unacceptable risks.

One solution to AI alignment proposed by Stuart Russell, a leading AI researcher, is the assistance game, also called a cooperativ

... (read more)
mic
18
0
0

Some quick thoughts:

  • Strong +1 to actually trying and not assuming a priori that you're not good enough.
  • If you're at all interested in empirical AI safety research, it's valuable to just try to get really good at machine learning research.
  • An IMO medalist or generic "super-genius" is not necessarily someone who would be a top-tier AI safety researcher, and vice versa.
  • For trying AI safety technical research, I'd strongly recommend How to pursue a career in technical AI alignment.
1
tlevin
Thanks for these points, especially the last one, which I've now added to the intro section.

As a countervailing perspective, Dan Hendrycks thinks that it would be valuable to have automated moral philosophy research assistance to "help us reduce risks of value lock-in by improving our moral precedents earlier rather than later" (though I don't know if he would endorse this project). Likewise, some AI alignment researchers think it would be valuable to have automated assistance with AI alignment research. If EAs could write a nice EA Forum post just by giving GPT-EA-Forum a nice prompt and revising the resulting post, that could help EAs save time... (read more)

Distributed computing seems to be a skill in high demand among AI safety organizations. Does anyone have recommendations for resources to learn about it? Would it look like using the PyTorch Distributed package or something like a microservices architecture?

2
Yonatan Cale
To answer your question: My best idea is working on a production distributed product that has bugs around that and debug them, and/or set up a system like that, perhaps with k8s. This answer is pretty bad, all these ideas are very hard and will take a long time to implement. My prior is that you might have a bottle neck which is easier to tackle than this one. Wanna share what your goal is, where you stand, and what you're planning to do?
mic
14
0
0

I feel somewhat concerned that after reading your repeated writing saying "use your AGI to (metaphorically) burn all GPUs", someone might actually do so, but of course their AGI isn't actually aligned or powerful enough to do so without causing catastrophic collateral damage. At least the suggestion encourages AI race dynamics – because if you don't make AGI first, someone else will try to burn all your GPUs! – and makes the AI safety community seem thoroughly supervillain-y.

Points 5 and 6 suggest that soon after someone develops AGI for the first time, th... (read more)

Quoting Scott Alexander here:

I agree it's not necessarily a good idea to go around founding the Let's Commit A Pivotal Act AI Company.

But I think there's room for subtlety somewhere like "Conditional on you being in a situation where you could take a pivotal act, which is a small and unusual fraction of world-branches, maybe you should take a pivotal act."

That is, if you are in a position where you have the option to build an AI capable of destroying all competing AI projects, the moment you notice this you should update heavily in favor of short timelines

... (read more)
5
Prometheus
If someone manages to create a powerful AGI, and the only cost for most humans is that it burns their GPUs, this seems like an easy tradeoff for me. It's not great, but it's mostly a negligible problem for our species. But I do agree using governance and monitoring is a possible option. I'm normally a hardline libertarian/anarchist, but I'm fine going full Orwellian in this domain.
3
aog
Strongly agreed. Somehow taking over the world and preventing anybody else from building AI seems like a core part of the plan for Yudkowsky and others. (When I asked about this on LW, somebody said they expected the first aligned AGI to implement global surveillance to prevent unaligned AGIs.) That sounds absolutely terrible -- see risks from stable totalitarianism.  If Yudkowsky is right and the only way to save the world is by global domination, then I think we're already doomed. But there's lots of cruxes to his worldview: short timelines, short takeoff speeds, the difficulty of the alignment problem, the idea that AGI will be a single entity rather than many different systems in different domains. Most people in AI safety are not nearly as pessimistic. I'd much rather bet on the wide range of scenarios where his dire predictions are incorrect. 

Thanks for writing this! I've seen Hilary Greaves' video on longtermism and cluelessness in a couple university group versions of the Intro EA Program (as part of the week on critiques and debates), so it's probably been influencing some people's views. I think this post is a valuable demonstration that we don't need to be completely clueless about the long-term impact of presentist interventions.

I'm really sorry that my comment was harsher than I intended. I think you've written a witty and incisive critique which raises some important points, but I had raised my standards since this was submitted to the Red Teaming Contest.

mic
12
0
0

For future submissions to the Red Teaming Contest, I'd like to see posts that are much more rigorously argued than this. I'm not concerned about whether the arguments are especially novel.

My understanding of the key claim of the post is, EA should consider reallocating some more resources from longtermist to neartermist causes. This seems plausible – perhaps some types of marginal longtermist donations are predictably ineffective, or it's bad if community members feel that longtermism unfairly has easier access to funding – but I didn't find the four reaso... (read more)

Thanks for the reply . Let me  just address the things I think are worth responding to.

For future submissions to the Red Teaming Contest, I'd like to see posts that are much more rigorously argued than this. I'm not concerned about whether the arguments are especially novel.

Ouch.  My humble suggestion: maybe be more friendly to outsiders, especially ones supportive and warm, when your movement has a reputation for being robotic/insular? Or just say "I don't want anyone who is not part of the movement to comment." Because that is the very obvious ... (read more)

Some quick thoughts:

  • EA Virtual Programs should be fine in my opinion, especially if you think you have more promising things to do than coordinating logistics for a program or facilitating cohorts
  • The virtual Intro EA Program only has discussions in English and Spanish. If group members would much prefer to have discussions in Hungarian instead, it might be useful for you to find some Hungarian-speaking facilitators.
  • Like Jaime commented, if you're delegating EA programs to EA Virtual Programs, it's best for you to have some contact with participants, especi
... (read more)
Answer by mic4
0
0

I see two new relevant roles on the 80,000 Hours job board right now:

Here's an excerpt from Anthropic's job posting. It's looking for basic familiarity with deep learning and mechanistic interpretability, but mostly nontechnical skills.

In this role you would:

  • Partner closely with the interpre
... (read more)

You might want to share this project idea in the Effective Environmentalism Slack, if you haven't already done so.

Is the application form "EAGxBerkeley, India & Future Forum Organizing Team Expression of Interest" supposed to have questions asking about whether you're interested in organizing the Future Forum? I don't see any; I only see questions about EAGxBerkeley and EAGxIndia.

2
Vaidehi Agarwalla 🔸
Yes i just added it - thanks for the flag! 
mic
10
0
0

From my experience with running EA at Georgia Tech, I think the main factors are:

  • not prioritizing high-impact causes
  • not being interested in changing their career plans
  • lack of high-impact career opportunities that fit their career interests, or not knowing about them
  • not having the skills to get high-impact internships or jobs
mic
16
0
0

I think I was primarily concerned that negative information about the campaign could get picked up by the media. Thinking it over now though, that motivation doesn't make sense for not posting about highly visible negative news coverage (which the media would have already been aware of) or not posting concerns on a less publicly visible EA platform, such as Slack. Other factors for why I didn't write up my concerns about Carrick's chances of being elected might have been that:

  • no other EAs seemed to be posting much negative information about the campaign, a
... (read more)

Thanks for the suggestion, just copied the critiques of the "especially useful" post over!

mic
12
0
0

Before the election was decided, I agreed with the overall point that donating, phone banking, or door-knocking for the campaign seemed quite valuable. At the same time, I want to mention a couple critiques I have (copied from my comment on "Some potential lessons from Carrick’s Congressional bid")

  • The post claims "The race seems to be quite tight. According to this poll, Carrick is in second place among likely Democratic voters by 4% (14% of voters favor Flynn, 18% favor Salinas), with a margin of error of +/- 4 percentage points." However, it declines to
... (read more)
mic
56
0
0

Overall, I agree with Habryka's comment that "negative evidence on the campaign would be 'systematically filtered out'". Although I maxed out donations to the primary campaign and phone banked a bit for the campaign, I had a number of concerns about the campaign that I never saw mentioned in EA spaces. However, I didn't want to raise these concerns for fear that this would negatively affect Carrick's chances of winning the election.

Now that Carrick's campaign is over, I feel more free to write my concerns. These included:

  • The vast majority of media coverage
... (read more)

I'd recommend cross-posting your critiques of the "especially useful" post onto that post — will make it easier for anyone who studies this campaign later (I expect many people will) to learn from you.

9
Aaron Gertler 🔸
Thanks for sharing all of this! I'm curious about your fear that these comments would negatively affect Carrick's chances. What was the mechanism you expected? The possibility of reduced donations/volunteering from people on the Forum? The media picking up on critical comments? If "reduced donations" were a factor, would you also be concerned about posting criticism of other causes you thought were important for the same reason?  I'm still working out what makes this campaign different from other causes (or maybe there really are similar issues across a bunch of causes).    One thing that comes to mind is time-sensitivity: if you rethink your views on a different cause later, you can encourage more donations to make up for a previous reduction. If you rethink views on a political campaign after Election Day, it's too late.  If that played a role, I can think of other situations that might exert the same pressure — for example, organizations running out of runway having a strong fundraising advantage if people are worried about dooming them. Not sure what to do about that, and would love to hear ideas (from anyone, this isn't specifically aimed at Michael).
mic
18
0
0

Another introductory post about why one may want to care about insect welfare: Does Insect Suffering Bug You? - Faunalytics (Jesse Gildesgame, 2016).

Recently, activists have started campaigning against silk because they believe the production process is cruel to silkworms. Many people respond to these campaigns with skepticism: who cares about silkworms? It’s easy to feel for the chinchillas, foxes, and other furry mammals used in fur clothing. But insects like silkworms are a harder sell. It seems crazy to grant moral consideration to a bug.

Nonetheless, t

... (read more)
mic
19
0
0

The Qualia Research Institute might be funding-constrained but it's questionable whether it's doing good work; for example, see this comment here about its Symmetry Theory of Valence.

3
Question Mark
Even if the Symmetry Theory of Valence turns out to be completely wrong, that doesn't mean that QRI will fail to gain any useful insight into the inner mechanics of consciousness. Andrew Zuckerman sent me this comment previously on QRI's pathway to impact, in response to Nuño Sempere's criticisms of QRI. The expected value of QRI's research may therefore have a very high degree of variance. It's possible that their research will amount to almost nothing, but it's also possible that their research could turn out to have a large impact. As far as I know, there aren't any other EA-aligned organizations that are doing the sort of consciousness research that QRI is doing.
Load more