All of Rubi J. Hudson's Comments + Replies

Does the LTFF ever counter-offer with an amount that would move the grant past the funding bar for cost-effectiveness? I would guess that some of these hypothetical applicants would accept a salary at 80% of what they applied for, and if the grants are already marginal then a 25% increase in cost-effectiveness could push them over the bar.

Yes, we do that fairly regularly. We haven't discussed people's counterfactual options above, but this often makes a 20% reduction in grant size unattractive.

Pronatalist.org is not an EA group. It's great that EA considerations have started entering the public consciousness and I would love if every charity was expected to answer "why is this the most effective thing you could be doing?", but that doesn't mean that any group claiming their mission is really important is part of EA. It's very difficult to argue a rigorous case that promoting pronatalist sentiment is an effective use of money or time, and so far they haven't.

Rather than ask how we can build more (and better) groups, ask whether we should.

Pronatalist.org is not an EA group. ... It's very difficult to argue a rigorous case that promoting pronatalist sentiment is an effective use of money or time, and so far they haven't.

The founders did write a detailed (and poorly received) post arguing for considering demographic collapse as a high-priority cause area.

Was Ben Pace shown these screenshots before he published his post?

Yes he was

Not all of them. He was shown the first set of screenshots, showing clearly that we said we were going to get her food. 

To be fair, we didn't show the screenshots about the mashed potatoes yet, which was proof we went out to get her vegan food. 

But we did show him there was vegan food in the house, which I think is an extremely important detail. 

Not feeding a sick friend (she was not working for us at the time) when she didn't have any food is cruel and uncaring. 

Offering to cook a sick friend vegan food in the house and her p... (read more)

With regards to #2, I shared your concern, and I thought Habryka's response didn't justify that the cost of a brief delay was sufficient if there was a realistic chance of evidence being provided to contradict the main point of this post.

However, upon reflection, I am skeptical that such evidence will be provided. Why did Nonlinear not provide at least some of the proof they claim to have, in order to justify time for a more comprehensive rebuttal? Or at least describe the form the proof will take? That should be possible, if they have specific evidence in... (read more)

Thanks for writing this up, I find quadratic funding falls into a class of mechanisms that are too clever by half in a way that makes them very fragile to the modelling assumptions. 

Also, I love how it's often posts with innocuous titles like 'Some thoughts on quadratic funding" that completely demolish an idea.

"People saying things that are mildly offensive but not worth risking an argument by calling out, and get tiring after repeated exposure" is just obviously a type of comment that exists, and is what most people mean when they say microaggression. Your paper debunking it alternates between much stricter definitions and claiming an absence of evidence for something that very clearly is going to be extremely hard to measure rigorously.

I'll edit to comment to note that you dispute it, but I stand by the comment. The AI system trained is only as safe as the mentor, so the system is only safe if the mentor knows what is safe. By "restrict", I meant for performance reasons, so that it's feasible to train and deploy in new environments.

Again, I like your work and would like to see more similar work from you and others. I am just disputing the way you summarized it in this post, because I think that portrayal makes its lack of splash in the alignment community a much stronger point against the community's epistemics than it deserves.

3
Michael_Cohen
1y
Thank you for the edit, and thank you again for your interest. I'm still not sure what you mean by a person "having access to the ground truth of the universe". There's just no sense I can think of where it is true that this a requirement for the mentor. "The system is only safe if the mentor knows what is safe." It's true that if the mentor kills everyone, then the combined mentor-agent system would kill everyone, but surely that fact doesn't weight against this proposal at all. In any case, more importantly a) the agent will not aim to kill everyone regardless of whether the mentor would (Corollary 14), which I think refutes your comment. And b) for no theorem in the paper does the mentor need to know what is safe; for Theorem 11 to be interesting, he just needs to act safely (an important difference for a concept so tricky to articulate!). But I decided these details were beside the point for this post, which is why I only cited Corollary 14 in the OP, not Theorem 11.

How does Rationalist Community Attention/Consensus compare? I'd like to mention a paper of mine published at the top AI theory conference which proves that when a certain parameter of a certain agent is set sufficiently high, the agent will not aim to kill everyone, while still achieving at least human-level intelligence. This follows from Corollary 14 and Corollary 6. I am quite sure most AI safety researchers would have confidently predicted no such theorems ever appearing in the academic literature. And yet there are no traces of any minds bei

... (read more)
9
Michael_Cohen
1y
Thanks! I should have clarified it has received some interest from some people. "When you can design and implement an agent that acts and updates its beliefs in a certain way and can restrict the initial beliefs to a set containing the desired ones". That is the "certain agent" I am talking about.  "Restrict" is an odd word choice, since the set can be as large as you like as long as it contains the truth. "and incorporate a human into the process who has access to the ground truth of the universe." This is incorrect; can I ask you to edit your comment? Absolutely nothing is assumed about the human mentor, certainly no access to the ground truth of the universe; it could be a two-year-old or a corpse!  That would just make the Mentor-Level Performance Corollary less impressive. I don't deny that certain choices about the agent's design make it intractable. This is why my main criticism was "People don't seem to bother investigating or discussing whether their concerns with the proposal are surmountable." Algorithm design for improved tractability is the bread and butter of computer science.

The standard academic failure mode is to make a number of assumptions for tractability that severely lower the relevance of the results (and the more pernicious failure mode is to hide those assumptions).

Perhaps, but at least these assumptions are stated. Most work leans on similarly strong assumptions (for tractability, brevity, or lack of rigour meaning you don't even realise you are doing it) but doesn't state them.

An EA steelman example of similar points of thinking are EAs who are incredibly anti-working for OpenAI or Deepmind at all because it safety washes and pushes capabilities anyways. The criticism here is the way EA views problems means EA will only go towards solution that are piecemeal rather than transformative. A lot of Marxists felt similarly to welfare reform in that it quelled the political will for "transformative" change to capitalism. 

For instance they would say a lot of companies are pursuing RLHF in AI Safety not because it's the correct way

... (read more)

Thanks for the communication, and especially giving percentages. Would you be able to either break it down by grants for individuals vs. grants to organizations, or note if the two groups were affected equally? While I appreciate knowing how high the bar has risen in general, I would be particularly interested in  how high it has risen for the kinds of applications I might submit in the future.

What are your intuitions regarding length? What's the minimum time needed for a fellowship to be impactful, and at what length does it hit diminishing returns?

5
Joel Becker
1y
I want to leave open the possibility that fellowships are not an effective thing to do regardless of their length, so maybe the minimum time is 0. But, conditional on thinking otherwise/going ahead with it... 1. I think it might be helpful to think about diminishing returns for participant stays, vs. fellowship length. You can imagine a 10 year-long program where people only stay for 4 months max at a time. This isn't as silly as it sounds -- very few people stay from the very beginning to very end of these things anyway, so program beginning/end don't seem like very important milestones to respect (vs. participant stays). 2. I think participant stays hit diminishing returns at 3 months. (Much more weakly held/pulled from nowhere than claims in this post.) 3. I worked ~24/7 as an organizer,  which became challenging after ~5 months.

Seems worth asking in interviews "I'm concerned about advancing capabilities and shortening timelines, what actions is your organization taking to prevent that", with the caveat that you will be BSed.

Bonus: You can turn down roles explicitly because they're doing capabilities work, which if it becomes a pattern may incentivize them to change their plan.

2
Yonatan Cale
2y
I agree, see foot note 2

This comment is object-level, perhaps nitpicky, and I quite like your post on a high level.

Saving a life via, say, malaria nets gets you two benefits:

1. The person saved doesn't die, meeting their preference for continuing to exist

2. The externalities of that person continuing to live, such as foregone grief by their family and community.

I don't think it's too controversial to say that the majority of the benefit from saving a life goes to the person whose life is saved, rather than the people who would be sad that they died. But the IDinsights survey only... (read more)

3
TheOtherHannah
2y
Yeah, and the IDinsight study only looked at #2 from your list above , which is one of the limitations and reasons more research would be good. This hits at a "collectivist culture vs individualist culture" nuance too, I suspect, because that could influence the weightings of #1 vs #2. In a 2012 blog post Holden wrote about the GiveWell approach being purposefully health and life-based as this is possibly the best way to give agency to distant communities: https://blog.givewell.org/2012/04/12/how-not-to-be-a-white-in-shining-armor/ And they also have a note somewhere on their website about flow-on effects: GiveWell assumes the flow on effects from giving health/life-saving interventions is probably more cost effective than flow on effects from infrastructural interventions which end up improving health and lifespan. In response to your comment about deferring to a hypothetical community who gives no life-saving intervention for people under 9 years old: if people had good access to information and resources, and their group decision was to focus a large amount of resource on saving lives of extremely old people on the community ... Maybe we should do this? I say this because I can think of reasons a community might want grandparents around for another few years (e.g. to pass on language, culture, knowledge) instead of more children at the moment. I think, if a community was at massive risk of loss of culture, the donors' insistence on saving young lives over the elders' lives could be incredibly frustrating. Not saying this to make any conclusions, but just as a counter-example that introduces a little more nuance than "morally wrong to let under 9yo's die unnecessarily."

There are lots of advantages to being based in the Bay Area. It seems both easier and higher upside to solve the Berkeley real estate issue that to coordinate a move away from the Bay Area.

2
BrownHairedEevee
8mo
Open Phil funds pro-housing advocacy, whose benefits are especially concentrated in areas like Berkeley, so these benefits will flow through to the EA and AIS communities as well.

I love the idea of a Library of EA! It would be helpful to eventually augment it with auxiliary and meta-information, probably through crowdsourcing among EAs. Each book could also be associated with short and medium summaries of the key arguments and takeaways, and warnings about which sections were later disproven or controversial (or a warning that the whole thing is a partial story/misleading). There's also a lot of overlap and superseding within the books (especially within the rationality and epistemology section), so it would be good to say "If you'... (read more)

4
RobBensinger
2y
My picks for a Core Longtermist EA Bookshelf (I don't see myself as having any expertise on what belongs in a Core Neartermist EA Bookshelf) would be: * HPMoR ↔ Scout Mindset * Rationailty: A-Z ↔ Good and Real * SSC (Abridged) * Superintelligence * Inadequate Equilibria ↔ Modern Principles of Economics (Cowen and Tabarrok) * Getting Things Done (Allen) Some people hate Eliezer's style, so I tried to think of books that might serve as replacements for at least some of the core content in RAZ etc. If I got a slightly longer list, I might add: How to Measure Anything, MPE, The Blank Slate (Pinker), Zero to One (Thiel), Focusing (Gendlin). Note that I tried to pick books based on what I'd expect to have a maximally positive impact if lots of people-who-might-help-save-the-future read them, not based on whether the books 'feel EA' or cover EA topics. Including R:AZ is sort of cheating, though, since it's more like six books in a trenchcoat and therefore uses up my Recommended EA Reading Slots all on its own. :p I haven't read the vast majority of books on the longer list, and if I did read them, I'd probably change my recommendations a bunch. I've read only part of The Blank Slate and Good and Real, and none of MPE, How to Measure Anything, or Focusing, so I'm including those partly on how strongly others have recommended them, and my abstract sense of the skills and knowledge the books impart.

In addition to EAG SF, there are some other major events and a general concentration of EAs happening in this 2-week time span in the Bay Area, so it might be generally good to come to the Bay around this time. 

Which other events are happening around that time? 

1
Dan Elton
2y
There's also a retreat being run for ACX and rationality meetup organizers (https://www.rationalitymeetups.org/new-page-4) July 21 - 24, and a lot of pre-events and after parties planned for EAG SF. (I can send people  GDocs with lists if anyone is interested.. I'm not sure if the organizers want them shared publicly). 
2
isaakfreeman
2y
A non-comprehensive list of events, to my knowledge: EAG SF, MLAB, many socials open to EAs (announced e.g. the Bay Area EA Facebook), apply-only or invite-only retreats, like GCP's X-Risk summits.  (I wouldn't be surprised if there were 2x more events than listed here.) I hope this is helpful.

their approaches are correlated with each other. They all relate to things like corrigibility, the current ML paradigm, IDA, and other approaches that e.g. Paul Christiano would be interested in.

You need to explain better how these approaches are correlated, and what an uncorrelated approach might look like. It seems to me that, for example, MIRI's agent foundations and Anthropic's prosaic interpretability approaches are wildly different!

By the time you get good enough to get a grant, you have to have spent a lot of time studying this stuff. Unpaid, mind y

... (read more)
1
NicholasKross
2y
Point 1: I said "Different from MIRI but correlated with each other". You're right that I should've done a better job of explaining that. Basically, "Yudkowksy approaches (MIRI) vs Christiano approaches (my incomplete read of most of the non-MIRI orgs). I concede 60% of this point. Point 2: !!! Big if true, thank you! I read most of johnswentworths' guide to being an independent researcher, and the discussion of grants was promising. I'm getting a visceral sense of this from seeing (and entering) more contests, bounties, prizes, etc. for alignment work. I'm working towards the day when I can 100% concede this point. (And, based on other feedback and encouragement I've gotten, that day is coming soon.)

I think your "digital people lead to AI" argument is spot on, and basically invalidates the entire approach. I think getting whole brain emulation working before AGI is such a longshot that the main effect of investing in it is advancing AI capabilities faster.

5
Steven Byrnes
2y
Strong agree, see for example my post Randal Koene on brain understanding before whole brain emulation
3
GMcGowan
2y
I definitely think it's an (the most?) important argument against. Some of this comes down to your views on timelines which I don't really want to litigate here.  I guess I don't know how much research leading to digital people is likely to advance AI capabilities. A lot of the early work was of course inspired by biology, but it seems like not much has come of it recently. And it seems to me that we can focus on the research needed to emulate the brain, and try not to understand it in too much detail.

Hopefully one day they grow big enough to hire an executive assistant.

While I'm familiar with literature on hiring, particularly unstructured interviews, I think EA organizations should give serious consideration to the possibility that they can do better than average. In particular, the literature is  correlational, not causal, with major selection biases, and is certainly not as broadly applicable as authors claim.

From Cowen and Gross's book Talent, which I think captures the point I'm trying to make well:
> Most importantly, many of the research studies pessimistic about interviewing focus on unstructured interview... (read more)

2
Joseph Lemien
2y
I think you are right. Like many things related to organizational behavior, I often think "in the kingdom of the blind the one-eyed man is king." So many organizations do such a poor job with hiring, even if we choose to merely adopt some evidence-based practices it can seem very impressive in comparison.

One of the keys things you hit on is  "Treating expenditure with the moral seriousness it deserves. Even offhand or joking comments that take a flippant attitude to spending will often be seen as in bad taste, and apt to turn people off."

However, I wouldn't characterize this as an easy win, even if it would be an unqualified positive. Calling out such comments when they appear is straightforward enough, but that's a slow process that could result in only minor reductions. I'd be interested in hearing ideas for how to change attitudes more thoroughly and quickly, because I'm drawing a blank.

I like many books on the list, but I think you're doing a disservice by trying to recommend too  many books at once. If you can cut it down to 2-3 in each category, that gives people a better starting point.

5
Aaron Gertler
2y
If you want recommendations, just take the first couple of  items in each category. They are rated in order of how good I think they are. (That's if you trust my taste — I think most people are better off just skimming the story summaries and picking up whatever sounds interesting to them.)

It's a bit surprising, but not THAT surprising. 50 more technical AI safety researchers would represent somewhere from a 50-100% increase  in the total number, which could be a justifiable use of 10% of OpenPhil's budget.

Great writeup! 

Is there an OpenPhil source for "OpenPhil values a switch to an AI safety research career as +$20M in expected value"? It would help me a lot in addressing some concerns that have been brought up in local group discussions.

6
Arjun Panickssery
2y
Update: I think he actually said "very good" AI safety researcher or something and I misremembered. The conversation was in January and before I knew anything much about the EAverse.
4
Arjun Panickssery
2y
This is what I remember Devansh (whom I pinged about your comment; I'll update when he replies) telling me when I first called him. I might have misremembered.
5
Chris Leong
2y
I’m surprised by that figure. $1billion would only lead on 50 AI safety researchers and they seem to only have any $10 billion.

Even before a cost-benefit analysis, I'd like to see an ordinal ranking of priorities. For organizations like the CEA,  what would they do with a 20% budget increase? What would they cut if they had to reduce their budget by 20%? Same thing for specific events, like EAGs. For a student campus club, what would they do with $500 in funding? $2,000? $10,000? I think this type of analysis would be helpful for determining if some of the spending that appears more frivolous is actually the least important.

To clear up my identity, I am not Seán and do not know him. I go by Rubi in real life, although it is a nickname rather than my given name. I did not mean for my account to be an anonymous throwaway, and I intend to keep on using this account on the EA Forum. I can understand how that would not be obvious as this was my first post, but that is coincidental. The original post generated a lot of controversy, which is why I saw it and decided to comment.

You spoke to 20+ reviewers, half of which were sought out to disagree with you, and not a single one could

... (read more)

Hi Carla,

Thanks for taking the time to engage with my reply. I'd like to engage with a few of the points you made.

First of all, my point prefaced with 'speaking abstractly' was genuinely that. I thought your paper was poorly argued, but certainly within acceptable limits that it should not result in withdrawn funding. On a sufficient timeframe, everybody will put out some duds, and your organizations certainly have a track record of producing excellent work. My point was about avoiding an overcorrection, where consistently low quality work is guaranteed so... (read more)

>it's important that we don't condition funding on agreement with the funders' views.

Surely we can condition funding on the quality of the researcher's past work though? Freedom of speech and freedom of research are both important, but taking a heterodox approach shouldn't guarantee a sinecure either. 

If you completely disagree that people consistently producing bad work should not be allocated scare funds, I'm not sure we can have a productive conversation.

If you completely disagree that people consistently producing bad work should not be allocated scare funds, I'm not sure we can have a productive conversation.

I theoretically agree, but I think it's hard to separate judgements about research quality from disagreement with its conclusions, or even unrelated opinions on the authors.

For example, I don't think the average research quality by non-tenured professors (who are supposedly judged by the merits of their work) is better than that of tenured professors.

That puts a huge and dictatorial responsibility on funders in ways that are exactly what the paper argued are inappropriate.


If not the funders,  do you believe anyone should be responsible for ensuring harmful and wrong ideas are not widely circulated? I can certainly see the case that even wrong, harmful ideas should only be addressed by counterargument. However, I'm not saying that resources should be spent censoring wrong ideas harmful to EA, just that resources should not be spent actively promoting them. Funding is a privilege, consistently makin... (read more)

-6
Guy Raveh
2y

I thought the paper itself was poorly argued, largely as a function of biting off too much at once. Several times the case against the TUA was not actually argued, merely asserted to exist along with one or two citations for which it is hard to evaluate if they represent a consensus. Then, while I thought the original description of TUA was accurate, the TUA response to criticisms was entirely ignored. Statements like "it is unclear why a precise slowing and speeding up of different technologies...across the world is more feasible or effective than the sim... (read more)

6
CarlaZoeC
2y
Saying thepaper is poorly argued is not particularly helpful or convincing. Could you highlight where and why Rubi? Breadth does not de-facto mean poorly argued.  If that was the case then most of the key texts in  x-risk would all be poorly argued. Importantly, breadth was necessary to make a critique. There are simply many interrelated matters that are worth critical analysis.  Several times the case against the TUA was not actually argued, merely asserted to exist along with one or two citations for which it is hard to evaluate if they represent a consensus. As  David highlights in his response: we do not argue against the TUA, but point out the unanswered questions we observed. We do not argue against the TUA , but highlight assumptions that may be incorrect or smuggle in values. Interestingly, it's hard to find  how you believe the piece is both polemic but also not directly critiquing the TUA sufficiently.  Those two criticisms are in tension.  If you check our references, you will see that we cite many published papers that treat common criticisms and open questions of the TUA (mostly by advancing the research).  You spoke to 20+ reviewers, half of which were sought out to disagree with you, and not a single one could provide a case for differential technology? Of course there are arguments for it, some of which are discussed in the forum. Our argument is that there is a lack of peer-review evidence to support differential technological development as a cornerstone of a policy approach to x-risk. Asking that we articulate and address every hypothetical counterargument is an incredibly high-bar, and one that is not applied to any other literature in the field (certainly not the key articles of the TUA we focus on). It would also make the paper far longer and broader. Again,  your points are in tension.  I think the paper would have been better served by focusing on a single section, leaving the rest to future work. The style of assertions rather than ar
[anonymous]2y52
0
0

I would agree that the article is too wide-ranging. There's a whole host of content ranging from criticisms of expected value theory, arguments for degrowth, arguments for democracy, and then criticisms of specific risk estimates. I agreed with some parts of the paper, but it is hard to engage with such a wide range of topics. 

Several times the case against the TUA was not actually argued


I think that they didn't try to oppose the TUA in the paper, or make the argument against it themselves. To quote: "We focus on the techno-utopian approach to existential risk for three reasons. First, it serves as an example of how moral values are embedded in the analysis of risks. Second, a critical perspective towards the techno-utopian approach allows us to trace how this meshing of moral values and scientific analysis in ERS can lead to conclusions, which, from a different perspective, loo... (read more)

Priors should matter! For example, early rationalists were (rightfully) criticized for being too open to arguments from white nationalists,  believing they should only look at the argument itself rather than the source. It isn't good epistemics to ignore the source of an argument and their potential biases (though it isn't good epistemics to dismiss them out of hand either based on that, of course).

6
anonymousEA
2y
I don't see a dichotomy between "ignoring the source of an argument and their potential biases" and downvoting a multi-paragraph comment on the grounds that it used less-than-charitable language about Silicone Valley billionaires. Based on your final line I'm not sure we disagree?