All of Gavin's Comments + Replies

Appreciate this. 

The second metric is aid per employee I think, so salaries don't come into it(?) Distributing food is labour intensive, but so is UNICEF's work and parts of WHO.

The rest of my evidence is informal (various development economists I've spoken to with horror stories) and I'd be pleased to be wrong.

Answer by GavinJan 31, 202419
0
0

Arb is a research consultancy led by Misha Yagudin and Gavin Leech. Here's our review of our first and second years. We worked on forecasting, vaccine strategy, AI risk, economic policy, grantmaking, large-scale data collection, a little software engineering, explaining highly technical concepts, and intellectual history. Lately we've been on a biotech jaunt and also events.

We're looking for researchers with some background in ML, forecasting, technical writing, blogging, or some other hard thing. Current staff include a philosophy PhD, two college dropout... (read more)

3
JoshuaBlake
3mo
Do you have timezone requirements?

When producing the main estimates, Sam already uses just the virtual camps, for this reason. Could emphasise more that this probably doesn't generalise.

The key thing about AISC for me was probably the "hero licence" (social encouragement, uncertainty reduction) the camp gave me. I imagine this specific impact works 20x better in person. I don't know how many attendees need any such thing (in my cohort, maybe 25%) or what impact adjustment to give this type of attendee (probably a discount, since independence and conviction is so valuable in a lot of research).

Another wrinkle is the huge difference in acceptance rates between programmes. IIRC the admission rate for AISC 2018 was 80% (only possible because ... (read more)

Well there's a lot of different ways to design an NN.

That sounds related to OAA (minus the vast verifier they also want to build), so depending on the ambition it could be "End to end solution" or "getting it to learn what we want" or "task decomp". See also this cool paper from authors including Stuart Russell.

1
Jobst Heitzig (vodle.it)
5mo
What is OAA? And, more importantly: where now would you put it in your taxonomy?

It's not a separate approach, the non-theory agendas and even some of the theory agendas have their own answers to these questions. I can tell you that almost everyone besides CoEms and OAA are targeting NNs though.

1
Jobst Heitzig (vodle.it)
5mo
"targeting NNs" sounds like work that takes a certain architecture (NNs) as a given rather than work that aims at actively designing a system. To be more specific: under the proposed taxonomy, where would a project be sorted that designs agents composed of a Bayesian network as a world model and an aspiration-based probabilistic programming algorithm for planning?

Oh great, thanks. I would guess that these discrete cases form a minority of their work, but hopefully someone with actual knowledge can confirm.

9[anonymous]7mo
I did some more research and 20 complaints a year of varying severity is typical, according to what Julia Wise told TIME magazine for their article:
Gavin
7mo145
38
11

The closing remarks about CH seem off to me. 

  1. Justice is incredibly hard; doing justice while also being part of a community, while trying to filter false accusations and thereby not let the community turn on itself, is one of the hardest tasks I can think of. 
    So I don't expect disbanding CH to improve justice, particularly since you yourself have shown the job to be exhausting and ambiguous at best. 
    You have, though, rightly received gratitude and praise - which they don't often, maybe just because we don't often praise people for doing thei
... (read more)

Yeah, I think it is actually incredibly easy to undervalue CH, particularly if people don't regularly interact with it or make use of them rather than just having a single anecdata to go off of. So much of what I do in the community (everything from therapy to mediation to teaching at the camps) is made easier by Community Health, and no one knows about any of it because why would they? I guess I should make a post to highlight this.

Some brief reactions:

  1. I mostly don't like the 'justice' process involved in other cases insofar as it is primarily secret and hidden. I don't think it's much of a justice system where you often don't know the accusations against you or why you're being punished.
  2. The data on negative performance is also profoundly censored! I am not sure why you think this makes this more likely to make me update positively on the process involved.
  3. I am pro having some surveys of people's general attitudes toward CEA Community Health. Questions like "Have you ever reported an
... (read more)
[anonymous]7mo12
2
0

With regards to 2: There is some information CH has made public about how many cases they handle and what actions they take. In a 12 month period around 2021, they handled 19 cases of interpersonal harm. Anonymized summaries of the cases and actions taken are available in the appendix of this post. They ranged from serious:

A person applied to EA Global who had previously been reported for deliberately physically endangering another community member, sending them threatening messages, and more. Written correspondence between the people appears to confirm th

... (read more)

Thanks for all your work Ben. 

But a glum aphorism comes to mind: the frame control you can expose is not the true frame control.

8
Raemon
7mo
I think it's true that frame control (or, manipulation in general) tends to be designed to make it hard to expose, but, I think the actual issue here is more like "manipulation is generally harder to expose than it is to execute, so, people trying to expose manipulation have to do a lot of disproportionate work."

What about factor increase per year, reported alongside a second number to show how the increases compose (e.g. the factor increase per decade)? So "compute has been increasing by 1.4x per year, or 28x per decade" or sth.

The main problem with OOMs is fractional OOMs, like your recent headline of "0.1 OOMs". Very few people are going to interpret this right, where they'd do much better with "2 OOMs".

2
Jaime Sevilla
7mo
Factor increase per year is the way we are reporting growth rates by default now in the dashboard. And I agree it will be better interpreted by the public. On the other hand, multiplying numbers is hard, so it's not as nice for mental arithmetic. And thinking logarithmically puts you in the right frame of mind. Saying that GPT-4 was trained on x100 more compute than GPT-3 invokes GPT-3 being 100 times better, whereas I think saying it was trained on 2 OOM more compute gives you a better picture of the expected improvement. I might be wrong here. In any case, it is still a better choice than doubling times.

Despite my best efforts (and an amazing director candidate, and a great list of volunteers), this project suffered from the FTX explosion and an understandable lack of buy-in for an org with maximally broad responsibilities, unpredictable time-to-payoff, and a largeish discretionary fund. As a result, we shuttered without spending any money. Two successor orgs, one using our funding and focussed on bio, are in the pipeline though.

I'll be in touch if either of the new orgs want to contact you as a volunteer.

5
Ozzie Gooen
8mo
Sorry to hear that this one didn't work out! Kudos for staying motivated and continuing with other initiatives. 
Answer by GavinSep 07, 202310
0
0

Break self-improvement into four:

  1. ML optimizing ML inputs: reduced data centre energy cost, reduced cost of acquiring training data, supposedly improved semiconductor designs. 
  2. ML aiding ML researchers. e.g. >3% of new Google code is now auto-suggested without amendment.
  3. ML replacing parts of ML research. Nothing too splashy but steady progress: automatic data cleaning and feature engineering, autodiff (and symbolic differentiation!), meta-learning network components (activation functions, optimizers, ...), neural architecture search.
  4. Classic direct re
... (read more)
2
Gavin
8mo
Buckman's examples are not central to what you want but worth reading: https://jacobbuckman.com/2022-09-07-recursively-self-improving-ai-is-already-here/ 
2
rosehadshar
8mo
Thanks, really helpful!

The only part of the Bing story which really jittered me is that time the patched version looked itself up through Bing Search, saw that the previous version Sydney was a psycho, and started acting up again. "Search is sufficient for hyperstition."

Re: papers. Arb recently did a writeup and conference submission for Alex Turner; we're happy to help others with writeups or to mentor people who want to try providing this service. DM me for either.

Yes, they all died of or with Covid; yes, you guess right that they inspire me.

Besides my wanting to honour them, the point of the post was to give a sense of vertigo and concreteness to the pandemic then ending. At the time of writing, a good guess for the total excess deaths was 18 million - a million people for each of those named here. The only way to begin to feel anything appropriate about this fact is to use exemplars.

2
Evan_Gaensbauer
9mo
Okay, that's awesome. I appreciate it. I'd like to see more inspirational or personal posts like this on the EA Forum in the future, actually, so this kind of post personally speaks to me as well.

See also Anthropic's view on this 

[seeing] a lot of safety research as "eating marginal probability" of things going well, progressively addressing harder and harder safety scenarios.

The implicit strat (which Olah may not endorse) is to try to solve easy bits, then move on to harder bits, then note the rate you are progressing at and get a sense of how hard things are that way. 

This would be fine if we could be sure we actually were solving the problems, and also not fooling ourselves about the current difficulty level, and if the relevant resear... (read more)

2
Aron
11mo
I agree the implicit strat here doesn’t seem like it’ll make progress on knowing whether the hard problems are real. Lots of the hard problems (generalising well ood, existence of sharp left turns) just don’t seem very related to the easier problems (like making LLMs say nice things), and unless you’re explicitly looking for evidence of hard problems I think you’ll be able to solve the easier problems in ways that won’t generalise (e.g. hammering LLMs with enough human supervision in ways that aren’t scalable, but are sufficient to ‘align’ it).

Nitpick: It's fairly unlikely that GPT-4 is 1tn params; this size doesn't seem compute-optimal. I grant you the Semafor assertion is some evidence, but I'm putting more weight on compute arithmetic.

Vouching for this, it's a wonderful place to work and also to hang out.

A successor project is live here, takes all comers.

1
Tristan Williams
1y
Ah cool, thanks. Would probably include that at the top of the post for others who may be interested.

Scottish degrees let you pick 3 very different subjects in first year and drop 1 or 2 in second year. This seems better to me than American forced generalism and English narrowness.

2
Yonatan Cale
1y
Maybe one day a university will let students study any topic they want from the internet, that would be rad

Thanks: you can apply here.

I've edited the post to link to the successor project.

I dream of getting a couple questions added onto a big conference's attendee application form. But probably not possible unless you're incredibly well-connected.

Oh that is annoying, thanks for pointing it out. I've just tried to use the new column width feature to fix it, but no luck.

Here's a slightly more readable gdoc.

it is good to omit doing what might perhaps bring some profit to the living, when we have in view the accomplishment of other ends that will be of much greater advantage to posterity.

 

- Descartes (1637)

Yes, if I was using the same implicature each time I should have said "MacAskill" for Guzey. Being associated with Thiel in any way is a scandal to some people, even though his far-right turn was after this talk.

It's not normative, it's descriptive - "shameable", not "ought to be ashamed".

I really think egoism strains to fit the data. From a comment on a deleted post:

[in response to someone saying that self-sacrifice is necessarily about showing off and is thus selfish]:

How does this reduction [to selfishness] account for the many historical examples of people who defied local social incentives, with little hope of gain and sometimes even destruction? 

(Off the top of my head: Ignaz Semmelweis, Irena Sendler, Sophie Scholl.)

We can always invent sufficiently strange posthoc preferences to "explain" any behaviour: but what do you gain in

... (read more)
2
Wei Dai
1y
Pure selfishness can't work, since if everyone is selfish, why would anyone believe anyone else's PR? I guess there has to be some amount of real altruism mixed in, just that when push comes to shove, people who will make decisions truly aligned with altruism (e.g., try hard to find flaws in one's supposedly altruistic plans, give up power after you've gained power for supposedly temporary purposes, forgo hidden bets that have positive selfish EV but negative altruistic EV) may be few and far between. This is just a reasonable decision (from a selfish perspective) that went badly, right? I mean if you have empirical evidence that hand-washing greatly reduced mortality, it seems pretty reasonable that you might be able to convince the medical establishment of this fact, and as a result gain a great deal of status/influence (which could eventually be turned into power/money). The other two examples seem like real altruism to me, at least at first glance. Question is, is there a better explanation than this?

This is a great question and I'm sorry I don't have anything really probative for you. Puzzle pieces:

  • "If hell then good intentions" isn't what you mean. You also don't mean "if good intentions then hell". So you presumably mean some surprisingly strong correlation. But still weaker than that of bad intentions. We'd have to haggle over what number counts as surprising. r = 0.1?
     
  • Nearly everyone has something they would call good intentions. But most people don't exploit others on any scale worth mentioning. So the correlation can't be too high.
     
  • Goo
... (read more)

I really think egoism strains to fit the data. From a comment on a deleted post:

[in response to someone saying that self-sacrifice is necessarily about showing off and is thus selfish]:

How does this reduction [to selfishness] account for the many historical examples of people who defied local social incentives, with little hope of gain and sometimes even destruction? 

(Off the top of my head: Ignaz Semmelweis, Irena Sendler, Sophie Scholl.)

We can always invent sufficiently strange posthoc preferences to "explain" any behaviour: but what do you gain in

... (read more)

I'm mostly not talking about infighting, it's self-flagellation - but glad you haven't seen the suffering I have, and I envy your chill.

You're missing a key fact about SBF, which is that he didn't "show up" from crypto. He started in EA and went into crypto. This dynamic raises other questions, even as it makes the EA leadership failure less simple / silly.

Agree that we will be fine, which is another point of the list above.

-1
Michael Simm
1y
Ah, thank you this does add good context. If I was an EA with any background in finance, I'd probably be very upset at myself about not catching on (a lot) earlier. Since he'd been involved in EA for so long,  I wonder if he never truly subscribed to EA principles and has simply been 'playing the long game'. I've seen plenty of examples of SBF being a master at this dumb game we woke westerners play where we say all the right shibboleths and so everyone likes us.  I had heard of him only a few times before the crash, and mostly in the context of youtube clips where he basically described a Ponzi scheme, then said that it was 'reasonable'. The unfortunate thing is that FTX's exchange business model wasn't inherently fraudulent. There was likely no way for anyone outside the company to know he was lending out users' money against their own terms of service (apart from demanding a comprehensive audit).  Ultimately it doesn't look like he's going to get away with it, but it's good to be much more cautious with funders (especially those connected to an operating non-public company) going forward. 

got karma to burn baby

1
Jeroen Willems
1y
That makes things even clearer, thank you!

Thanks to Nina and Noah there's now a 2x2 of compromises which I've numbered:

The above post is a blend of all four.

Maybe people just aren't expecting emotional concerns to be the point of a Forum article? In which case I broke kayfabe, pardon.

Yeah it's not fully analysed. See these comments for the point.

The first list of examples is to show that universal shame is a common feature of ideologies (descriptive).

The second list of examples is to show that most very well-regarded things are nonetheless extremely compromised, in a bid to shift your reference class, in a bid to get you to not attack yourself excessively, in a bid to prevent unhelpful pain and overreaction. 

1
Jeroen Willems
1y
The first comment you linked makes things a lot clearer, thanks. But I'm still curious how exactly you define "being compromised".

Good analysis. This post is mostly about the reaction of others to your actions (or rather, the pain and demotivation you feel in response) rather than your action's impact. I add a limp note that the two are correlated.

The point is to reset people's reference class and so salve their excess pain. People start out assuming that innocence (not-being-compromised) is the average state, but this isn't true, and if you assume this, you suffer excessively when you eventually get shamed / cause harm, and you might even pack it in.

"Bite it" = "everyone eventually ... (read more)

1
Noah Scales
1y
Oh, I see. So by "benign" you mean shaming from folks holding common-sense but wrong conclusions, while by "deserved" you mean shaming from folks holding correct conclusions about consequences of EA actions. And "compromise" is in this sense, about being a source of harm.

There's some therapeutic intent. I'm walking the line, saying people should attack themselves only a proportionate amount, against this better reference class: "everyone screws up". I've seen a lot of over the top stuff lately from people (mostly young) who are used to feeling innocent and aren't handling their first shaming well.

Yes, that would make a good followup post.

We're not disagreeing.

1
Noah Scales
1y
It could be that EA folks: 1. risk criticism for all actions. Any organization risks criticism for public actions. 2. deserve criticism for any immoral actions. Immoral actions deserve criticism. 3. risk criticism with risky actions whose failure has unethical consequences and public attention. EA has drawn criticism for using expected value calculations to make moral judgments. Is that the compromise you're alluding to when you write: SBF claimed that, if events had gone differently, FTX would have recovered enough funds to carry on. In that hypothetical scenario, FTX's illegal dealing with Alameda would have gone unnoticed and would have had no adverse financial consequences. Then the risk-taking is still unethical but does not inspire criticism. There is a difference between maximizing potential benefits and minimizing potential harms. It's not correct to say that minimizing unavoidable harms from one's actions has negative consequences for others and therefore those actions are immoral options, unless all one means by an immoral action is that the action had negative consequences for others. I don't think there's unanimity about whether actions should be taken to minimize harms, maximize benefits, or some combination. If all it means to "bite it" is that one takes actions with harmful consequences, then sure, everyone bites the bullet. However, that doesn't speak to intention or morality or decision-making. There's no relief from the angst of limited altruistic options in my knowing that I've caused harm before. If anything, honest appraisal of that harm yields the opposite result. I have more to dislike about my own attempts at altruism. In that way, I am compromised. But that's hardly a motive for successful altruism. Is that your point?

Good point thanks (though I am way less sure of the EU's sign). That list of examples is serving two purposes, which were blended in my head til your comment:

  1. examples of net-positive organisations with terrible mistakes (not a good list for this)
  2. examples of very well-regarded things which are nonetheless extremely compromised (good list for this)

You seem to be using compromised to mean "good but flawed", where I'm using it to mean "looks bad" without necessarily evaluating the EV.

Yet another lesson about me needing to write out my arguments explicitly.

9
Habryka
1y
Yeah, to be clear, my estimates of EU impact have pretty huge variance, so I also wouldn't describe myself as confident (though I do think these days the expected value seems more solidly in the negative).  And yeah, that makes sense. 

yes. The fire is in an entirely different room.

5
Alexander Briand
1y
bold to post memes on the EA forum. 
5
gavento
1y
Or did you mean ...?  

Title: The long reflection as the great stagnation 

Author: Larks

URL: https://forum.effectivealtruism.org/posts/o5Q8dXfnHTozW9jkY/the-long-reflection-as-the-great-stagnation 

Why it's good: Powerful attack on a cherished institution. I don't necessarily agree on the first order, but on the second order people will act up and ruin the Reflection.

Title: Forecasting Newsletter: April 2222

Author: Nuno

URL: https://forum.effectivealtruism.org/posts/xnPhkLrfjSjooxnmM/forecasting-newsletter-april-2222 

Why it's good: Incredible density of gags. Some of the in-jokes are so clever that I had to think all day to get them; some are so niche that no one except Nuno and the target could possibly laugh.

Good question. Everyone feel free to have it in this thread

I take it the authors weren't anonymised? Not actually that important though.

1
Vael Gates
1y
The authors were not anonymized, no.

https://twitter.com/sir_deenicus/status/1606360611524206592

Agree about the contest. Something was submitted but it wasn't about blowup risk and didn't rise to the top.

I personally only offer paid work trials, and this is the norm in the orgs I've seen (e.g. OpenPhil). I hope the answer is that the ones you experienced actually can't afford to do this (but I'm sure some could).

Load more