All of calebp's Comments + Replies

I could see people upvoting this post because they think it should be more like -10 than -21. I personally don't see it as concerning that it's "only" on -21.

6
Jason
8d
That's plausible, although I find it somewhat less likely to be a complete or primary explanation based on my recollection of voting patterns/trends on past eugenics-adjacent posts.  In any event, I don't think "was down-voted a lot" (from harfe's comment) would be a complete summary of the voting activity.

Sorry, it wasn't clear. The reference class I had in mind was cause prio focussed resources on the EA forum.

I think people/orgs do some amount of this, but it's kind of a pain to share them publicly. I prefer to share this kind of stuff with specific people in Google Docs, in in-person conversations, or on Slack.

I also worry somewhat about people deferring to random cause prio posts, and I'd guess that on the current margin, more cause prio posts that are around the current median in quality make the situation worse rather than better (though I could see it going either way).

2
Jamie_Harris
13d
Thanks! When you say "median in quality" what's the dataset/category that you're referring to? Is it e.g. the 3 ranked lists I referred to, or something like "anyone who gives this a go privately"?

Thanks for writing this.

I disagree with quite a few points in the total utilitarianism section, but zooming out slightly, I think that total utilitarians should generally still support alignment work (and potentially an AI pause/slow down) to preserve option value. If it turns out that AIs are moral patients and that it would be good for them to spread into the universe optimising for values that don't look particularly human, we can still (in principle) do that. This is compatible with thinking that alignment from a total utilitarian perspective is ~neutral - but it's not clear that you agree with this from the post.

6
Matthew_Barnett
16d
I think the problem with this framing is that it privileges a particular way of thinking about option value that prioritizes the values of the human species in a way I find arbitrary. In my opinion, the choice before the current generation is not whether to delay replacement by a different form of life, but rather to choose our method of replacement: we can either die from old age over decades and be replaced by the next generation of humans, or we can develop advanced AI and risk being replaced by them, but also potentially live much longer and empower our current generation's values. Deciding to delay AI is not a neutral choice. It only really looks like we're preserving option value in the first case if you think there's something great about the values of the human species. But then if you think that the human species is special, I think these arguments are adequately considered in the first and second sections of my post.

I'm interested in examples of this if you have them.

2
trevor1
1mo
Yeah, a lot of them are not openly advertised for good reasons. One example that's probably fine to talk about is NunoSempere's claim that EAforum is shifting towards catering to new or marginal users.  The direct consequence is reducing the net quality of content on EAforum, but it also allows it to steer people towards events as they get more interested in various EA topics, where they can talk more freely without worrying about saying things controversial, or get involved directly with people working on those areas via face-to-face interaction. And it doesn't stop EAforum from remaining a great bulletin board for orgs to publish papers and updates and get feedback. But at first glance, catering towards marginal users normally makes you think that they're just trying to do classic user retention. That's not what's happening; this is not a normal forum and that's the wrong way to think about it.

Oh, I thought you might have suggested the live thing before, my mistake. Maybe I should have just given the 90-day figure above.

(That approach seems reasonable to me)

I answered the first questions above in an edit of the original comment. I’m pretty sure when I re-ran the analysis with decided in last 30 days it didn’t change the results significantly (though I’ll try and recheck this later this week - in our current setup it’s a bit more complicated to work out than the stats I gave above).

I also checked to make sure that only looking at resolved applications and only looking at open applications didn’t make a large difference to the numbers I gave above (in general, the differences were 0-10 days).

2
Elizabeth
2mo
I'm not following- what does it mean to say you've calculated resolution time to applications that haven't been resolved?

Oh, right - I was counting "never receiving a decision but letting us know" as a decision. In this case, the number we'd give is days until the application was withdrawn.

We don't track the reason for withdrawals in our KPIs, but I am pretty sure that process length is a reason for a withdrawal 0-5% of the time.

I might be missing why this is important, I would have thought that if we were making an error it would overestimate those times - not underestimate them.

4
Rebecca
1mo
My point was that if someone withdraws their application because you were taking so long to get back to them, and you count that as the date you gave them your decision, you’re artificially lowering the average time-till-decision metric. Actually the reason I asked if you’d factored in withdrawn application not how was to make sure my criticism was relevant before bringing it up - but that probably made the criticism less clear

I'm not sure sorry, I don't have that stat in front of me. I may be able to find it in a few days.

Empirically, I don't think that this has happened very much. We have a "withdrawn by applicant status", which would include this, but the status is very rarely used.

In any case, the numbers above will factor those applications in, but I would guess that if we didn't, the numbers would decrease by less than a day.

6
Rebecca
2mo
My point is more around the fact that if a person withdraws their application, then they never received a decision and so the time till decision is unknown/infinite, it’s not the time until they withdrew.

We do have a few processes that are designed to do this (some of which are doing some of the things you mentioned above). Most of the long delays are fairly uncorrelated (e.g. complicated legal issue, a bug in our application tracker ...).

I think it could be good to put these number on our site. I liked your past suggestion of having live data, though it's a bit technically challenging to implement - but the obvious MVP (as you point out) is to have a bunch of stats on our site. I'll make a note to add some stats (though maintaining this kind of information can be quite costly, so I don't want to commit to doing this).

In the meantime, here are a few numbers that I quickly put together (across all of our funds).

Grant decision turnaround times (mean, median):

  • applied in the last 30 days = 14 d
... (read more)
9
Jeff Kaufman
1mo
How are these included? Is it that in you count ones that haven't closed as if they had closed today? (A really rough way of dealing with this would be to count ones that haven't closed as if they will close in as many days from now as they've been open so far, on the assumption that you're on average halfway through their open lifetime.)
4
Rebecca
1mo
Is the repetition of “applied in the last 30 days” possibly a typo?

Thanks for engaging with my criticism in a positive way.

Regarding how timely the data ought to be, I don't think live data is necessary at all - it would be sufficient in my view to post updated information every year or two.

I don't think "applied in the last 30 days" is quite the right reference class, however, because by-definition, the averages will ignore all applications that have been waiting for over one month. I think the most useful kind of statistics would:

  1. Restrict to applications from n to n+m months ago, where n>=3
  2. Make a note of what percent
... (read more)

Do you know what proportion of applicants fill out the feedback form?

Is there (or might it be worthwhile for there to be) a business process to identify aged applications and review them at intervals to make sure they are not "stuck" and that the applicant is being kept up to date? Perhaps "aged" in this context would operationalize as ~2x the median decision time and/or ~>90-95th percentile of wait times? Maybe someone looks at the aged list every ~2 weeks, makes sure the application isn't "stuck" in a reasonably fixable way, and reviews the last correspondence to/from the applicant to make sure their information about timeframes is not outdated?

2
Elizabeth
2mo
what does 30/60/90 days mean? Grants applied to in the last N days? Grants decided on in the last N?  How do the numbers differ for acceptances and rejections?  What percent of decisions (especially acceptances) were made within the timeline given on the website?  Can you share more about the anonymous survey? How has the satisfaction varied over time? 
4
Rebecca
2mo
Are you factoring in people who withdraw their application because of how long the process was taking?

If it's okay with you, I'd prefer not to have screenshots of my emails posted right now; I'm happy to rethink this in a few days when it feels a bit lower pressure.

I generally don't write emails, assuming that they will be posted to a public place like the forum.

I believe our application form says

The Animal Welfare Fund, Long-Term Future Fund and EA Infrastructure Fund aim to respond to all applications in 2 months and most applications in 3 weeks. However, due to an unprecedentedly high load, we are currently unable to achieve our desired speedy turnarounds. If you need to hear back sooner (e.g., within a few weeks), you can let us know in the application form, and we will see what we can do. Please note that: EA Funds is low on capacity and may not be able to get back to you by either your stated deadline or t

... (read more)

One thing to note is that at the end of January, we rejected the original grant (which I believed that we wouldn't be able to show a clear public benefit for), and then said we were interested in a different version of the grant that seemed more defensible to me (subject to legal review). Since then, we have been working out whether we can make this alternate grant.

I didn't realise that Igor stopped taking clients completely, and I regret that I didn't make a stronger effort to understand the consequences of the unclear situation whilst we tried to understand the legal implications of making the grant.

I’m very sorry that you had such a bad experience here. Whilst I would disagree with some of the details here I do think that our communication was worse than I would have liked and I am very sorry for any hardship that you experienced. It sounds like a stressful process which could have been made much better if we had communicated more often and more quickly.

In my last email (March 4th), I said that we were exploring making this grant, but it’s legally challenging. Grants for mental health support are complicated, in general, as we have to show that there... (read more)

5
Igor Ivanov
2mo
Thank you for the attitude you expressed here, although I believe that you promised me to get back in a certain time. A part from your email from December 5 We sincerely regret the delay and assure you that we will provide you with an update within the next few days. Your email from December 18 We expect to give a decision this week. I provide parts of your emails because you expressed that you would rather not share everything in your other comments here.

One thing to note is that at the end of January, we rejected the original grant (which I believed that we wouldn't be able to show a clear public benefit for), and then said we were interested in a different version of the grant that seemed more defensible to me (subject to legal review). Since then, we have been working out whether we can make this alternate grant.

I didn't realise that Igor stopped taking clients completely, and I regret that I didn't make a stronger effort to understand the consequences of the unclear situation whilst we tried to understand the legal implications of making the grant.

If the funders get a nontrivial portion of the impact for early-stage projects then I think the AWF (inc. its donors) is very plausible.

4
Ben_West
2mo
Yeah, I am not sure how to treat meta. In addition to funders, Charity Entrepreneurship probably gets substantial credit for SWP, etc.

Here are some very brief takes on the CCM web app now that RP has had a chance to iron out any initial bugs. I'm happy to elaborate more on any of these comments.

  • Some praise
    • This is an extremely ambitious project, and it's very surprising that this is the first unified model of this type I've seen (though I'm sure various people have their own private models).
      • I have a bunch of quantitative models on cause prio sub-questions, but I don't like to share these publicly because of the amount of context that's required to interpret them (and because the methodolo
... (read more)
5
Derek Shiller
2mo
Thanks for recording these thoughts! Here are a few responses to the criticisms. This is a fair criticism: we started this project with the plan of providing somewhat authoritative numbers but discovered this to be more difficult than we initially expected and instead opted to express significant skepticism about the default choices. Where there was controversy (for instance, in how many years forward we should look), we opted for middle-of-the-road choices. I agree that it would add a lot of value to get reasonable and well-thought-out defaults. Maybe the best way to approach controversy would be to opt for different sets of parameter defaults that users could toggle between based on what different people in the community think. The ability to try to represent digital people with populations per star was a last-minute choice. We originally just aimed for that parameter to represent human populations. (It isn’t even completely obvious to me that stars are the limiting factor on the number of digital people.) However, I also think these things don’t matter since the main aim of the project isn’t really affected by exactly how valuable x-risk projects are in expectation. If you think there may be large populations, the model is going to imply incredibly high rates of return on extinction risk work. Whether those are the obvious choice or not depends not on exactly how high the return, but on how you feel about the risk, and the risks won't change with massively higher populations. If you think we’ll likely have an aligned super-intelligence within 100 years, then you might try to model this by setting risks very low after the next century and treating your project as just a small boost on its eventual discovery. However, you might not think that either superaligned AI or extinction is inevitable. One thing we don’t try to do is model trajectory changes, and those seem potentially hugely significant, but also rather difficult to model with any degree of confidence.

I don't understand the point about the complexity of value being greater than the complexity of suffering (or disvalue). Can you possibly motivate the intuition here? It seems to me like I can reverse the complex valuable things that you name, and I get their "suffering equivalents" e.g. (e.g. friendship -> hostility, happiness -> sadness, love -> hate ... etc.), and they don't feel significantly less complicated. 

I don't know exactly what it means for these things to be less complex; I'm imagining something like writing a Python program that simulates the behaviour of two robots in a way that is recognisable to many people as "friends" or "enemies" and measuring at the length of the program.

8
MichaelStJules
2mo
It's not that there aren't similarly complex reverses, it's that there's a type of bad that basically everyone agrees can be extremely bad, i.e. extreme suffering, and there's no (or much less) consensus on a good with similar complexity and that can be as good as extreme suffering is bad. For example, many would discount pleasure/joy on the basis of false beliefs, like being happy that your partner loves you when they actually don't, whether because they just happen not to love you and are deceiving you, or because they're a simulation with no feelings at all. Extreme suffering wouldn't get discounted (much) if it were based on inaccurate beliefs. A torturous solipsistic experience machine is very bad, but a happy solipsistic experience machine might not be very good at all, if people's desires aren't actually being satisfied and they're only deceived into believing they are.

Oli’s comment so people don’t need to click through

I thought some about the AI Safety camp for the LTFF. I mostly evaluated the research leads they listed and the resulting teams directly, for the upcoming program (which was I think the virtual one in 2023).

I felt unexcited about almost all the research directions and research leads, and the camp seemed like it was aspiring to be more focused on the research lead structure than past camps, which increased the weight I was assigning to my evaluation of those research directions. I considered for a while to

... (read more)

I agree that people in existing EA hubs are more likely to come across others doing high value work than people located outside of hubs.

That said, on the current margin, I still think many counterfactually connections happen at office spaces in existing EA hubs. In the context of non residential spaces, I’m not really sure who would use an EA office space outside existing EA hubs so I’m finding the comparison between office in a hub vs office outside a hub a little confusing (whereas with CEEALAR I understand who would use it).

1
Siao Si
3mo
I imagine there could be a useful office in a city with ~20 people using it regularly and ~100 people interested enough in EA to come to some events, and I wouldn't think of that city as an "EA hub". I also think eg. a London office has much more value than an eg. an Oxford or Cambridge office (although I understand all three to be hubs), even though Oxford and Cambridge have a higher EA-density.

I go back and forth on this. Sometimes, I feel like we are funding too many underperforming projects, but then some marginal project surprises me by doing quite well, and I feel better about the hits-based strategy. Over the last three months, we have moved towards funding things that we feel more confident in, mostly due to funding constraints.

I don't think that I have a great list of common communicable lessons, some high-level thoughts/updates that jump to mind:

  • in general, people will be worse than they expect when working in areas they have little e
... (read more)

I think the performance/talent of grantees and context is extremely important. 

That said, some programs that I am excited about that I believe many EAs are a good fit for:

  • University EA groups, particularly ones at top universities
  • Field-specific retreats/workshops/boot camps etc.
  • Career advising calls and other career-focused content
  • Writing high-quality blog posts

Some projects I have seen work well in the past, but I think they are a bad fit for most people:

  • Youtube channels
  • Mass media comms (like writing an op-ed in a popular newspaper)

Most of my views o... (read more)

(note that I'm not speaking about CEEALAR or any other specific EAIF applicants/grantees specifically)

I understand that CEEALAR has created a low-cost hotel/coworking space in the UK for relatively junior people to stay while they work on research projects relevant to GCRs. I think that you had some strategic updates recently so some of my impression of your work may be out of date. Supporting people early on in their impact-focused careers seems really valuable, I've seen lots of people go through in-person retreats and quickly start doing valuable work.

A... (read more)

3
Siao Si
3mo
I don't understand this consideration. It seems to me that people located in a place with a more robust existing community are the people that would counterfactually benefit the least from a place to interact with other EAs, because they have plenty of opportunities to do so already. I'm assuming by "hub" you mean "EA hub", but if by "hub" you mean "a place with high population density/otherwise a lot of people to talk to", then this makes sense. (Full disclosure: I was a grantee of CEEALAR last year; but I'm thinking about this in the context of non-residential office/co-working spaces like Meridian Office).

Doctors in the UK (like the ones that set this up) earn way less that $350k a year in general. Junior doctors (which are the majority of the UK doctor workforce) are very poorly paid, I think many of my friends made something like £14/hour for the first few years after qualifying.

2
Larks
4mo
According to the first result in Google, doctors' total pay, while significantly lower than the UK, is still significantly above the UK average, even for junior doctors. Their hourly rate is surprisingly low but that's mainly because they work very long hours.
4
Jason
4mo
This is also generally true of US trainees (residents and fellows). For the non-US people, there's 4 years of undergrad, 4 years of medical school, generally 3-5 years of residency, then an optional 1-3? years of fellowship for certain specialties. On the other hand, I think Lilly is correct insofar as 1 percent would be a pretty meager ask for most US attendings, even in lower-paying specialties.
6
lilly
4mo
I earn about $15/hour and donate much more than 1%. I don't think it's that hard to do this, and it seems weird to set such a low bar.

I didn't say that AI was software by definition - I just linked to some (brief definitions) to show that your claim afaict is not widely understood in technical circles (which contradicts your post). I don't think that the process of using Photoshop to edit a photo is itself a program or data (in the typical sense), so it seems fine to say that it's not software.

Definition make claims about what is common between some set of objects. It's fine for single members of some class to be different from every other class member. AI does have a LOT of basic stuff ... (read more)

5
Davidmanheim
4mo
One the first point, I think most technical people would agree with the claim: "AI is a very different type of thing that qualifies as software given a broad definition, but that's not how to think about it." And given that, I'm saying that we don't say " a videoconference meeting is different to other kinds of software in important ways," or "photography is different to other kinds of software in important ways" because we think of those as a different thing, where the fact that it's run on software is incidental. And my claim is that we should be doing that with AI.

I'll probably ask some of my ML engineer friends this week, but I am fairly sure that most ML people would be fine with calling AI products, models, etc. software. I don't have much of an opinion on whether calling AI systems software creates confusion or misunderstandings - I'd guess that calling AI software within policy circles is generally helpful (maybe you have a better alternative name).

Encyclopedia Britannica

Software, instructions that tell a computer what to do. Software comprises the entire set of programs, procedures, and routines associated w

... (read more)
9
abramdemski
4mo
I think these definitions are good enough to explain why AI models should not be classified as software: software is instructions that tell a computer what to do. Or a "program". While deep learning "weights" do tell a computer what to do (a model can be "run on some input" much like a computer program can), these weights do not resemble instructions/programs. 

Appealing to definitions seems like a bad way to argue about whether the conceptual model is useful or not. The operation of a computer system and the "software" used for  digital photography, or videoconferencing, or essay writing, is not typically considered software.  Do you think those should be called software, given that they fit into the definitions given?

I'm claiming that AI is distinct in many ways from everything else we typically think of as software, not that it doesn't fit a poorly scoped definition. Amd the examples of "collection[s... (read more)

Thanks for pointing that out. I re-read the post and now think that the OP was more reasonable. I'm sorry I missed that in the first place. I also didn't convey the more important message of "thank you for critiquing large, thorny, and important conclusions". Thinking about P(bio x-risk) is really quite hard relative to lots of research reports posted on the forum, and this kind of work seems important.

I don't care about the use of Bayesian language (or at least I think that bit you quoted does all the Bayesian language stuff I care about).

Maybe I should r... (read more)

I think reasonable even if EA Funds was its own org.

I was a bit disappointed by this post. I think I am sympathetic to the overall take and I’m a bit frustrated that many EAs are promoting or directly working on biorisk without imo compelling reports to suggest a relatively high chance of x-risk.

That said, this post seems to basically make the same error, it says that Ord’s estimates are extremely high but doesn’t really justify that claim or suggest a different estimate. It would be much more reasonable imo to say “Ord’s estimate is much higher than my own prior, and I didn’t see enough evidence to justify such a large update”.

8
JoshuaBlake
4mo
Except the use of Bayesian language, how is that different to the following passage?
6
David Thorstad
4mo
Thanks Caleb! I give reasons for skepticism about high levels of existential biorisk in Parts 9-11 of this series.

In fairness, I was prevented from posting a bunch of stuff and spent a long time (like tens of hours) workshopping text until legal council were happy with it. At least in one case I didn’t end up posting the thing because it didn’t feel useful after the various edits and it had been by then a long time since the event the post was about.

I think in hindsight the response (with the information I think the board had) was probably reasonable - but if similar actions were to be taken by EV when writing a post about Anthropic I’d be pretty upset about that. I wouldn’t use the word censoring in the real ftx case - but idk in the fictional Anthropic case I might?

0
Larks
4mo
I think it's worth not entangling the word 'censorship' with whether it is justified. During the Second World War the UK engaged in a lot of censorship, to maintain domestic morale and to prevent the enemy from getting access to information, but this seems to me to have been quite justified, because the moral imperative for defeating Germany was so great. Similarly, it seems quite possible to me that in the future CEA might be quite justified in instituting AI-related censorship, preventing people from publishing writing that disagrees with the house line. It seems possible to me that the FTX and EV related censorship was justified, though it is hard to tell, given that EV have never really explained their reasons, and I think the policy certainly had very significant costs. In the wake of FTX's collapse there was a lot of soul-searching and thinking about how to continue in the EA community and we were deprived of input from many of the best informed and most thoughtful people. My guess is this censorship was especially onerous on more junior employees for whom it was harder to justify the attorney review time, leading to a default answer of 'no'. So the reason I mentioned it wasn't that censorship is always a bad choice, or that, conditional on censorship being imposed, it is likely to be a mistake, given the situation. The argument is that who your leader is changes the nature of the situation, changing whether or not censorship is required, and the nature of that censorship. As an analogy, if Helen knew what was going to come, I imagine she might have written that report quite differently - with good reason. A hypothetical alternative CSET with a different leader would not have face such pressures.
4
Neel Nanda
4mo
Reasonable because you were all the same org, or reasonable even if EA Funds was its own org

I'm not sure that I would use the word censoring, but there were strict policies around what kinds of communications various EV orgs could do around FTX for quite a long time (though I don't think they were particularly unusual for an organisation of EVs size in a similar legal situation).

The wording of what Larks said makes it seem like over a number of years staff were prevented from expressing their true opinions on central EA topics

EV was fine with me publishing this. My experience was that it was kind of annoying to publish FTX stuff because you had to get review first, but I can't recall an instance of being prevented from saying something.

"Aggressively censored its staff" doesn't reflect my experience, but maybe reflects others', not sure.

calebp
4mo64
10
0
6
1

I'm excited that Zach is stepping into this role. Zach seems substantially better than my expectations for the new CEA CEO, and I expect the CEO hiring committee + Ben + the EV board had a lot to do with that (and probably lots of other people at CEA that I don't know about)!

Most CEA users and EA community members probably don't know Zach, so I thought it would be helpful to share some of my thoughts on them and this position (though I don't know Zach especially well, and these are just quick subjective takes). Thanks to @Ben_West for the nudge to do this.... (read more)

3
Geoffrey Miller
4mo
Caleb - thanks for this helpful introduction to Zach's talents, qualifications, and background -- very useful for those of us who don't know him! I agree that EA organizations should try very hard to avoid entanglements with AI companies such as Anthropic - however well-intentioned they seem. We need to be able to raise genuine concerns about AI risks without feeling beholden to AI corporate interests.

Zach is on Anthropic's Long-Term Benefit Trust. It's not super clear what this means, particularly in light of recent events with the Open AI board, but I am a bit concerned about the way that EA views Anthropic, and that the CEO of CEA being affiliated with Anthropic could make it more difficult for people within EA to speak out against Anthropic.

This is a very interesting point given that it seems that Helen's milquetoast criticism of OpenAI was going to be used as leverage to kick her off the OpenAI board, and that historically EV has aggressively censored its staff on important topics.

That would help me! Right now I mostly ignore the expertise/interest fields, but I could imagine using this feature to book 1:1s if people used a convention like the one you suggested.

+1 to the EAG expertise stuff, though I think that it’s generally just an honest mistake/conflicting expectations, as opposed to people exaggerating or being misleading. There aren’t concrete criteria for what to list as expertise so I often feel confused about what to put down.

 

@Eli_Nathan maybe you could add some concrete criteria on swapcard?

e.g. expertise = I could enter roles in this specialty now and could answer questions of curious newcomers (or currently work in this area)

interest = I am either actively learning about this area, or have invested at least 20 hours learning/working in this area .

4
Ivan Burduk
4mo
Hi Caleb, Ivan from the EAG team here — I'm responsible for a bunch of the systems we use at our events (including Swapcard). Thanks for flagging this! It's useful to hear that this could do with more clarity. Unfortunately, there isn't a way we can add help text or sub text to the Swapcard fields due to Swapcard limitations. However, we could rename the labels/field names to make this clearer..? For example * Areas of Expertise (3+ months work experience) * Areas of Interest (actively seeking to learn more) Does that sound like something that would be helpful for you to know what to put down? I'll take this to the EAG team and see if we can come up with something better. Let me know if you have other suggestions!
Answer by calebpDec 17, 202314
3
1

I think the closest thing to an EA perspective written relatively recently that is all in a single doc is probably this pdf of Holdens most important century sequence on cold takes.

I feel quite confused about what empirical evidence from the forum would change your mind. I personally don't think that the comment folding of weakly negative comments has much effect on future engagement (particularly if it's a reasonable comment).

I initially thought that if your hypothesis about hiding influencing voting is correct and people regularly do the voting on expected Karma, then I think you should expect to see quite a few "reasonable to you" low negative karma comments (as they have been voted negative and then stayed negative due to hiding ... (read more)

5
Jason
4mo
My view could be disconfirmed by various sorts of A/B testing. The most obvious of which would involve manipulating the amount of karma displayed to viewers shortly after posting. Let's say +3 of true karma for one group (A), true karma for a second group (B), and and -3 of true karma for the third (C). If my view that early negative votes can have a disproportionate effect is true, then we should see more downvotes and/or fewer upvotes coming from group C than groups A or B. Over a sufficient number of comments, this should be powered enough to reach statistical significance if there is a meaningful effect. This is similar to the methodology used by Muchnik et al. in their study on social influence bias on an unnamed upvoting/downvoting site.  Notably, Muchnik et al. observed significant differences even though their karma manipulation was limited to +1/-1 at the time of karma creation. I suggested +3/-3 here because the types of voters I think likely to employ a preemptive strategy would typically have +2/-2 weight on ordinary votes, and there could be more than one person employing the strategy. A less direct study design would be to show some users the comments as currently displayed (A), while other users see something akin to Reddit "contest mode" (B: random order, scores hidden). As one might guess, this strategy is used by some subreddits to mitigate the trend of users disproportionately upvoting comments that already had upvotes. For instance, on one major subreddit [link contains a few curse words], the top comment used to be on average posted 4.47 minutes after the post was made. Using contest mode in the first hour increased that to ~11 minutes, and the length of the top comment to almost double. This design couldn't prove my view, but a finding of no difference between A and B in a sufficiently powered sample would render it rather unlikely.  One might reasonably counter that Forum users are less likely to be swayed by others' votes than users on a l

I think downvoting with the expectation of upvotes is very reasonable. It's probably a better norm than voting based on current karma when only some people follow the vote based on current karma level policy and you care about the steady state karma behaviour.

7
Jason
4mo
The main reason the pattern Peter described is weird is that the mystery downvoter, who is drastically outnumbered by mystery upvoters, usually beats all of them to the ballot box. I would be more favorably inclined toward your theory if it were applied when a few upvotes were already on board. A very early downvote can have a disproportionate effect on a comment's visibility in some cases, or potentially influence other voters too much in a bandwagon effect of sorts. That's why, e.g., Reddit has a "contest mode" available, in which comments are randomly sorted and karma not displayed.

Nevertheless, my impression is that GiveWell's cost-effectiveness estimates are pretty close to encompassing all their thinking. Elie mentioned on the Clearer Thinking podcast that:

Fwiw I feel quite confused about how different GiveWell's recommendations would be if they were solely optimising for cost-effectiveness, I have heard different versions of how much they are optimising for this already based on different people that I speak to (and my impression is that most public materials do not say they are solely optimising for this).

Do you have a cost

... (read more)
6
GiveWell
4mo
Hi Vasco and Caleb, we appreciate the interest in the Global Health and Development Fund! This is Isabel Arjmand responding on behalf of GiveWell. We're grateful for the opportunity to manage this fund, and we think it's a great opportunity for donors who want to support highly cost-effective global health and development programs. We're also interested in having more in-depth conversations with Caleb and others involved in EA Funds about what the future of this fund should look like, and we’ll reach out to schedule that. In the meantime, here are some notes on our grantmaking and how donations to the fund are currently used. * We expect the impact of giving to the Global Health and Development Fund (GHDF) is about the same as giving to GiveWell's All Grants Fund: both go to the most impactful opportunities we've identified (across programs and organizations), and are a good fit for donors who'd like to support the full range of our grantmaking, including higher-risk grants and research. The online description of GHDF was written before the All Grants Fund existed (it launched in 2022), and the two funds are now filling a very similar niche. Caleb, we'd love to collaborate on updating the GHDF webpage to both reflect the existence of the All Grants Fund and include more recent grant payout reports. * In the broadest sense, GiveWell aims to maximize impact per dollar. Cost-effectiveness is the primary driver of our grantmaking decisions. But, “overall estimated cost-effectiveness of a grant” isn't the same thing as “output of cost-effectiveness analysis spreadsheet.” (This blog post is old and not entirely reflective of our current approach, but it covers a similar topic.) * The numerical cost-effectiveness estimate in the spreadsheet is nearly always the most important factor in our recommendations, but not the only factor. That is, we don’t solely rely on our spreadsheet-based analysis of cost-effectiveness when making grants.  * We don't have an instituti

Hi Vasco, quickly responding to a few of your points here

Let its donors know that donating to GHDF in its current form has a similar effect to donating to AGF (if that is in fact the case), instead of just describing GHDF as a “higher-risk” “higher-reward” alternative (to TCF). “Donating to this fund [GHDF] is valuable because it helps demonstrate to GiveWell that there is donor demand for higher-risk, higher-reward global health and development giving opportunities”. 

  • I do think that the GHDF is more committed to doing something like maximising the im
... (read more)
2
Jason
4mo
This implies a crux to me. Presumably the people running these charities seek funding from EA sources, despite knowing that counterfactually the bulk of that money would otherwise go to AGF/GHDF/et al. Do you think they disagree with your assessment of their effectiveness, perhaps due to different moral weights? I'm not suggesting your assessment is wrong -- my own tentative view is that there aren't (m)any places with room for large amounts of funding that would beat GiveWell AGF or similar on pure QUALYs (or equivalent). That's a second crux, I think. While that is an important purpose, it is not as predominant a purpose in my view. There are lots of monies out there that are practically restricted in a way that precludes AGF et al. from competing for them. This could be due to pre-defined government/foundation grant areas, or due to an individual donor's personal preferences, or a non-EA donor's desire for some higher quantum of warm fuzzies than organizations like AMF can provide. If we don't have anything to offer in those areas, we are conceding them to less effective charities. All funding niches have low-hanging fruit and monies that are harder to acquire. For instance, speaking from personal experience, there are a lot of US evangelical Christians who won't donate to an organization unless it is Christian-flavored enough. As you might guess, I do not share that view -- but it is what it is, and some people with these views are wealthy. Kaleem's recent quick take on zakat provides another possible example. Other donors really want to donate to mental-health causes, etc. Meanwhile, most GiveWell-recommended charities are significantly operating in a specific zone of fundraising (i.e., donors open to all sorts of charities without any self-imposed limitations). Thus, while it likely isn't cost-effective to fund organizations that operate in niches merely for the QUALYs they produce, it may be beneficial to support them in their earlier stages until they are
6
Vasco Grilo
4mo
Thanks for the clarifying comments, Caleb! I very much agree cost-effectiveness is all that matters, in the sense that one would ideally maximise the benefits for a given amount of resources. Of course, as you point out, this does not mean the cost-effectiveness number outputted by the spreadsheet is all that matters! It is often hard to formalise all factors which influence cost-effectiveness, and so the actual cost-effectiveness estimates one obtains are not everything. Nevertheless, my impression is that GiveWell's cost-effectiveness estimates are pretty close to encompassing all their thinking. Elie mentioned on the Clearer Thinking podcast that: This is in line with what you said about Open Phil and GiveWell agreeing with you. Thanks for questioning! I have clarified my reasons a little more updating the 1st 2 bullets of the section Case for donating to Giving What We Can's Global Health and Wellbeing Fund to: From your comment, it sounds like you have some concerns about GiveWell's recommendation process, which I think would be worth expanding on more publicly. I guess a public call might be helpful to find such people. Rethink Priorities might be open to doing some evaluations? They have been commisioned by GiveWell, and have experience incubating new projects. Do you have a cost-effectiveness bar as a fraction of the cost-effectiveness of GiveDirectly? It may be better to be explicit about it. GiveWell's is 10, and I believe Open Phil's is 20. Fair points. Maybe such charities will eventually (not initially) go on to use funds which would otherwise have gone to less effective charities?

Probably most controversially, while I view existential risk reduction work as tremendously important, I don’t donate any of the 10% of my income dedicated to effective charity in order to support this work. (I do view it as a critical global priority, which is why the vast majority of my time and effort are spent on it!) Principally, my lack of donations is because I don’t view the cause area as a charitable endeavor, rather than rational self-interest for myself and my family, which has obvious benefits to the broader world. This does not make it less i

... (read more)
4
Davidmanheim
4mo
Thanks for engaging. Despite the fact that I don't expect this to sound convincing or fully lay out the case for anything, I'll provide a brief partial sketch of some of my tentative views in order to provide at least a degree of transparency around my reasoning - and I'm unlikely to expand on this much futher here, as I think this takes a lot of time and intensive discussion to transmit, and I have other things I should be working on. First, I think you're accidentally assuming utilitarianism in my decisionmaking. I view my charitable giving from more of a contractualist and rights view, where the deontological requirement to impartially benefit others is only one aspect of my decisionmaking. This goes pretty deep into fundamental views and arguments which I think are going to be hard to explain quickly, or at all. Second, my initial commitment to charity, the one that led my to commit to giving 10% of my income, was to benefit the poor impartially, as one of several goals I had - embracing what Richard Chappell has called a benificentist view. I see the money I am putting aside as some degree of stewardship of money I have allocated to charity, which is given on behalf of others. Given that commitment, to the extent that I have goals which differ from benefitting the poor, there is a very, very high bar for me to abrogate that commitment and take that money to do other things. At the very least, having come to a personal conclusion that I care about the future and view existential risk as a priority does not do enough to override that commitment. As an aside, I'll note that it is rational for those with fewer opportunities and less access to capital to make shorter term decisions; my prioritization of my children and grandchildren is in large part because I'm comparably rich. And so the refusal to reallocate my giving of money commited to others has a lot to do with respecting preferences, even when I think they are "mistaken" - because the bar for overrising o

I think David means "giving motivated by impartiality" instead of giving to places that themselves are "cause neutral".

2
Vasco Grilo
4mo
Thanks for the comment, Caleb. That might be the case. On the other hand, the sentence I quoted seems to suggest people at GiveWell are "cause-neutral experts", and I think this applies more to people at Rethink Priorities, which works across multiple areas, and has a Worldview Investigation Team.

I'm not too worried about this kind of moral uncertainty. I think that moral uncertainty is mostly action-relevant when one moral view is particularly 'grabby' or the methodology you use to analyse an intervention seems to favour one view over another unfairly.

In both cases, I think the actual reason for concern is quite slippery and difficult for me to articulate well (which normally means that I don't understand it well). I tend to think that the best policy is to maximise the expected outcomes of the overall decision-making policy (which involves paying... (read more)

Could you give me some examples of these kinds of projects? I think, as Linch said, Manifund is probably their best bet, or posting on the EA forum asking for funding from individual donors.

Answer by calebpDec 09, 20236
1
0

I'd like to hear his advice for smart undergrads who want to build their own similarly deep models in important areas which haven't been thought about very much e.g. take-off speeds, the influence of pre-AGI systems on the economy, the moral value of insects, preparing for digital minds (ideally including specific exercises/topics/reading/etc.).

I'm particularly interested in how he formed good economic intuitions, as they seem to come up a lot in his thinking/writing.

Thanks! To be clear, this is a 'plan' instead of something we are 100% committed to delivering on in the way it's presented below. I think there are some updates to be made here, but I would feel bad if you made large irreversible decisions based on this post. We will almost certainly have a more official announcement if we do decide to commit to this plan.

3
James Herbert
5mo
Thanks for making that clear! 

I agree with the overall point, though I am not I've seen much empirical evidence for the GHD as a good starting point claim (or at least I think it's often overstated). I got into EA stuff though GHD, but, this may have just been because there were a lot more GHD/EA intro materials at the time. I think that the eco-system is now a lot more developed and I wouldn't be surprised if GHD didn't have much of an edge over cause first outreach (for AW or x-risk).

Maybe our analysis should be focussed on EA principles, but the interventions themselves can be bran... (read more)

Load more