All of Buhl's Comments + Replies

Thank you! 

Worth noting that our input was also very unevenly distributed – our original idea list includes ~40% AI-related ideas, ~15% bio, ~25% movement building / community infrastructure, and only ~20% other. (this was mainly due to us having better access to AI-related project ideas via our networks). If you’re interested in pursuing biosecurity- or movement building-related projects, feel free to get in touch and I can share some of our additional ideas – for the other areas I think we don’t necessarily have great ideas.

Thanks, appreciate your comment and the compliment!

On your questions:

2. The research process does consider cost-effectiveness as a key factor – e.g., the weighted factor model we used included both an “impact potential” and a “cost” item, so projects were favoured if they had high estimated impact potential and/or a low estimated cost. “Impact potential” here means “impact with really successful (~90th percentile) execution” – we’re focusing on the extreme rather than the average case because we expect most of our expected impact to come from tail outcomes... (read more)

The quick explanation is that I don't want people to over-anchor on it, given that the inputs are extremely uncertain, and that I think that a ranked list produced by a relatively well-respected research organisation is the kind of thing people could very easily over-anchor on, even if you caveat it heavily

(I'm in a similar position to Amber: Limited background (technical or otherwise) in AI safety and just trying to make sense of things by discussing them.)

Re: "I think you need to say more about what the system is being trained for (and how we train it for that). Just saying "facts about humans are in the data" doesn't provide a causal mechanism by which the AI acts in human-like ways, any more than "facts about clouds are in the data" provides a mechanism by which the AI role-plays being a cloud."

The (main) training process for LLMs is exactly to predict h... (read more)

6
Lucius Bushnaq
1y
"Could reasonably be described" is the problem here. You likely need very high precision to get this right. Relatively small divergences from human goals in terms of bits altered suffice to make a thing that is functionally utterly inhuman in its desires. This is a kind of precision that current AI builders absolutely do not have. Worse than that, if you train an AI to do a thing, in the sense of setting a loss function where doing that thing gets a good score on the function, and not doing that thing gets a bad score, you do not, in general, get out an AI that wants to do that thing. One of the strongest loss signals that trains your human brain is probably "successfully predict the next sensory stimulus". Yet humans don't generally go around thinking "Oh boy, I sure love successfully predicting visual and auditory data, it's so great." Our goals have some connection to that loss signal, e.g. I suspect it might be a big part of what makes us like art. But the connection is weird and indirect and strange.  If you were an alien engineer sitting down to write that loss function for humans, you probably wouldn't predict that they'd end up wanting to make and listen to audio data that sounds like Beethoven's music, or image data that looks like van Gogh's paintings. Unless you knew some math that tells you what kind of AI with what kind of goals g you get if you train on a loss function L over a dataset D. The problem is that we do not have that math. Our understanding of what sort of thinky-thing with what goals comes out at the end of training is close to zero. We know it can score high on the loss function in training, and that's basically it. We don't know how it scores high. We don't know why it "wants" to score high, if it's the kind of AI that can be usefully said to "want" anything. Which we can't tell if it is either. With the bluntness of the tools we currently possess, the goals that any AGI we make right now would have would effectively be a random draw
Buhl
1y14
0
1

Thank you for the important post!
 

“we might question how well neuron counts predict overall information-processing capacity”

My naive prediction would be that many other factors predicting information-processing capacity (e.g., number of connections, conduction velocity, and refractory period) are positively correlated with neuron count, such that neuron count is pretty strongly correlated with information processing even if it only plays a minor part in causing more information processing to happen. 

You cite one paper (Chitka 2009) that provides ... (read more)

Curious what you're referring to here and if there's any publicly available information about it? Couldn't find anything in ALLFEDs 2020 and 2021 updates. (I'm trying to estimate the cost-effectiveness of this kind of project as part of my work at Rethink Priorities)

Another failure mode I couldn’t easily fit into the taxonomy that might warrant a new category:

Competency failures - EAs are just ineffective at achieving things in the world due to lack of skills (eg comms, politics, org running) or bad judgement. Maybe this could be classed as a resource failure (for failing to attract people with certain skills) or a rigor failure (for failing to develop them/learn from others). Will try to think of a title beginning with R…

Minor points:

  • I was also considering something like value failures (EAs have the wrong moral the
... (read more)

Curious what people think of the argument that, given that people in the EA community have different rankings of the top causes, a close-to-optimal community outcome could be reached if individuals argmax using their own ranking?

(At least assuming that the number of people who rank a certain cause as the top one is proportional to how likely it is to be the top one.)

Buhl
2y22
0
0

[Shortform version of this comment here.]

Update: I helped Linch collect data on the undergrad degrees of exceptionally successful people (using some of the ex post metrics Linch mentioned).

Of the 32 Turing Award winners in the last 20 years, 6 attended a top 10 US university, 16 attended another US university, 3 attended Oxbridge, and 7 attended other non-US universities. (full data)

Of the 97 Decacorn company founders I could find education data for, 19 attended a top 10 US university, 32 attended another US university, and 46 attended non-US universities ... (read more)

1
Aaron Bergman
2y
Wow awesome analysis! Haven't read super carefully just yet but does seem to point in the opposite direction of the evidence I present insofar as it relates to uni CB. I think this would be worth making a normal high-level post! Blame it on me and link this comment if anyone disagrees lol My only super brief thought, which I don't think is controversial but I'll make explicit, is that reverse causality (i.e. Stanford signal->success metric)  is one reason you might not want to take this as knockdown evidence in favor of "Ivies select for talent," though I really don't have a great guess as to how much this is indeed the case
Buhl
2y15
0
0

Tl;dr: Most Turing Award winners and Decacorn company founders (i.e., exceptionally successful people) don’t attend US top universities, but there’s a fair amount of concentration.

In response to the post Most Ivy-smart students aren't at Ivy-tier schools and as a follow-up to Linch’s comment tallying the educational background of Field Medalists, I collected some data on the undergrad degrees of exceptionally successful people (using some of the  (imperfect) ex post metrics suggested by Linch).

Of the 32 Turing Award winners in the last 20 years, ... (read more)

Buhl
2y10
0
0

Thought-provoking post, thanks a lot for writing it! 

I broadly agree that it’s good for community builders to spend significant time on learning/direct work, especially if their long-term plan is not to do community building, but I think I disagree with some of your specific reasons.

I think the post sometimes conflates two senses of marketing. One is “pure” marketing, the other is marketing as you define it (i.e., marketing and ops), which includes things like organising content-heavy events and programs like fellowships. My instinct is that:

A. Most o... (read more)

No worries!

I don’t have strong opinions on a 4-week fellowship, no! I think my quick take would be that (a) it’s harder to teach the core EA ideas well in 4x1.5h sessions, (b) it’s harder to create a social community/have people become friends in 4 weeks, and (c) the group of people who’d commit to a 4-week program but not an 8-week program is relatively small, at least in a university group context. But I’m not too sure about this. It also seems plausible to me that 4 weeks could be better in  contexts like professional or city groups.

I’d be excited ... (read more)

4
JWehner
2y
Hey I'm one of the organisers of the PISE Fellowship and would like to weigh in on some of the points you made: (a) I agree that it's hard to cover all core ideas of EA well in 4 weeks. For example we were not able to fit in animal welfare.  So for a 4 week model it seems essential to offer things like discussion groups, in depth fellowships, etc. so people can keep learning after the fellowship is over.  (b) From my experience friendships and social engagement come more from social activities or working together than from a fellowship (might be different for others).  Again here it seems essential when running a 4 week fellowship to offer other ways of socialising and engaging. PISE does this by organising big social events and  by recruiting people into commitees after the fellowship ends. (c) Anecdotally, multiple people mentioned that they felt like 4 weeks was not a big commitment and joined because of that. I hope we soon have some data that can shed some light on this question. There will be a longer post about our experience with the 4 week fellowship soon. In the newly founded EA Delft we are planning to employ a different model. We will first run a 4 week intro fellowship (and advertise it as 4 weeks),  but then throughout offer people to continue the fellowship for another 4 weeks. This way the people only willing to join for 4 weeks will join, but  the ones willing to do the full 8 weeks will get to dive into more topics. We will share our experience and the results we get with this method at some point. 

Thanks for writing this post! I especially like the concrete alternatives with thoughtful upsides/downsides. As some others have said, I’d guess some of the downsides to the alternatives are quite significant, but would still love to see trials and to chat to anyone who runs trials.
 

A potentially useful alternative approach (especially for larger groups who can run multiple programs) is to have several alternative intro funnels at once. I.e. run the IF but also have a clear alternative for keen people with more background knowledge or who can quickly ... (read more)

Thanks for raising these points! A few of my (personal) reactions:

1. We definitely didn't intend for the post to presuppose that democracy is good for the long term. It’s true that most of the potential effects we identity are positive-leaning – but none of these effects, nor the all-things-considered effect, is a settled case. 

2. I think the question of what conditions allowed EA to come into existence is interesting, although not sure if that's the main positive impact of liberal democracy (especially given we don’t have super strong evidence that l... (read more)