All of Max Clarke's Comments + Replies

Just posting my reactions to reading this:

I find that rates are fairly high:

  • 25% of signatories have been accused of financial misconduct, and 10% convicted

That's really high?? Oh - this is not the giving what we can pledge๐Ÿ˜…

I estimate that Giving Pledgers are not less likely, and possibly more likely, to commit financial crimes than YCombinator entrepreneurs.

At what stage of YC? I guess that will be answered later. EDIT:

I previously estimated that 1-2% of YCombinator-backed companies with valuations over $100M had serious allegations of fraud.

... (read more)

Soaking screams food poisoning to me; especially with unclean water. Perhaps this is not a risk if done right, but this could be why it's not done.

6
NickLaing
8mo
Thanks Max, this was mentioned by be Eli as well but I don't think it's an issue. Because of the 1 hour of boiling, everything that was alive is very dead. The mechanism would have to be a toxin. There's no evidence I can find of toxin production after short term soaking.

Definitely, for example if people are bikeshedding (vigorously discussing something that doesn't matter very much)

Another proposal: Visibility karma remains 1 to 1, and agreement karma acts as a weak multiplier when either positive or negative.

So:

  • A comment with [ +100 | 0 ] would have a weight of 100
  • A comment with [ +100 | 0 ] but with 50โœ… and 50โŒ would have a weight of 100 + log10(50 + 50) = 200
  • A comment with [ +100 | 100โœ… ] would have a weight of say 100 * log10(โœ“100) = 200
  • A comment with [+0 | 1000โœ… ] would have a weight of 0.

Could also give karma on that basis.

However thinking about it, I think the result would be people would start using the visibility vote to express opinion even more...

Would you gift your karma if that option was available?

2
Nathan Young
1y
Yes. I think so, haven't looked at the utility curves but I imagine I can find people I think are underrated.
3
Michael_PJ
1y
Destroying it seems better. Gifting it requires identification of a worthy recipient and seems like it opens all kinds of additional problems.

This is good for calibrating what the votes mean across the responses

A little ambiguous between  "disagree karma & upvote karma should have equal weight" and "karma should have equal weight between people"

I think because the sorting is solely on karma, the line is "Everything above this is worth considering" / "Everything below this is not important" as opposed to "Everything above this is worth doing"

It's karma - which is kind of wrong here.

2
Nathan Young
1y
Can an opinion be right but unimportant?

One situation I use strong votes for is whenever I do "upvote/disagree" or "downvote/agree". I do this to offset others who tend not to split their votes.

I think some kind of "strong vote income", perhaps just a daily limit as you say, would work.

People who read this far seem to have upvoted

I would have expected the opposite corner of the two axis voting (because I think people don't like the language)

seems he has ended up giving more to the democratic party than ea lol

There seems to be two different conceptual models for AI risk.

The first is a model like in his report "Existential risk from power-seeking AI", in which he lays out a number of things, which, if they happen, will cause AI takeover.

The second is a model (which stems from Yudkowsky & Bosteom, and more recently in Michael Cohen's work https://www.lesswrong.com/posts/XtBJTFszs8oP3vXic/?commentId=yqm7fHaf2qmhCRiNA ) where we should expect takeover by malign AGI by default, unless certain things happen.

I personally think the second model is much more reasonable. Do you have any rebuttal?

5
Steven Byrnes
1y
See also Nate Soares arguing against Joeโ€™s conjunctive breakdown of risk here, and me here.

Likewise, I have a post from January suggesting that crypto assets are over-represented in the EA funding portfolio.

Probably the number of people actually pushing the frontier of alignment is more like 30, and for capabilities maybe 3000. If the 270 remaining alignment people can influence those 3000 (biiiig if, but) then the odds aren't that bad

Not sure what Rob is referring to but there are a fair few examples of org/people's purposes slipping from alignment to capabilities, eg. OpenAI

I myself find it surprisingly difficult to focus on ideas that are robustly beneficial to alignment but not to capabilities.

(E.g. I have a bunch of interpretability ideas. But interpretability can only have no impact on, or accelerate timelines)

Do you know if any of the alignment orgs have some kind of alignment research NDA, with a panel to allow any alignment-only ideas be public, but keep the maybe-capabilities ideas private?

4
Emrik
1y
Do you mean you find it hard to avoid thinking about capabilities research or hard to avoid sharing it? It seems reasonable to me that you'd actually want to try to advance the capabilities frontier, to yourself, privately, so you're better able to understand the system you're trying to align, and also you can better predict what's likely to be dangerous.

I think probably this post should be edited and "focus on low risk interventions first" put in bold in the first sentence and put right next to the pictures. Because the most careless people (possibly like me...) are the ones that will read that and not read the current caveats

2
Emrik
1y
You'd be well able to compute the risk on your own, however, if you seriously considered doing any big outreach efforts. I think people should still have a large prior on action for anything that looks promising to them. : )

An addendum is then:

  1. If Buying time interventions are conjunctive (ie. one can cancel out the effect of the others); but technical alignment is disjunctive

  2. If the distribution of people performing both kinds of intervention is mostly towards the lower end of thoughtfulness/competence, (which we should imo expect)

Then technical alignment is a better recommendation for most people.

In fact it suggests that the graph in the post should be reversed (but the axis at the bottom should be social competence rather than technical competence)

Rob puts it well in his comment as "social coordination". If someone tries "buying time" interventions and fails, I think that because of largely social effects, poorly done "buying time" interventions have potential to both fail at buying time and preclude further coordination with mainstream ML. So net negative effect.

On the other hand, technical alignment does not have this risk.

I agree that technical alignment has the risk of accelerating timelines though.

But if someone tries technical alignment and fails to produce results, that has no impact compare... (read more)

2
Max Clarke
1y
An addendum is then: 1. If Buying time interventions are conjunctive (ie. one can cancel out the effect of the others); but technical alignment is disjunctive 2. If the distribution of people performing both kinds of intervention is mostly towards the lower end of thoughtfulness/competence, (which we should imo expect) Then technical alignment is a better recommendation for most people. In fact it suggests that the graph in the post should be reversed (but the axis at the bottom should be social competence rather than technical competence)

I see!  Yes, I agree that more public "buying time" interventions (e.g. outreach) could be net negative. However, for the average person entering AI safety, I think there are less risky "buying time" interventions that are more useful than technical alignment.

That's a reasonable point - the way this would reflect in the above graph is then wider uncertainty around technical alignment at the high end of researcher ability

I would push back a little, the main thing is that buying time interventions obviously have significant sign uncertainty. Eg. your graph on median researcher "buying time" vs technical alignment, I think should have very wide error at the low end of "buying time", going significantly below 0 within the 95% confidence interval. Technical alignment is lots less risky to that extent.

9
Miranda_Zhang
1y
To clarify, you think that "buying time" might have a negative impact [on timelines/safety]? Even if you think that, I think I'm pretty uncertain of the impact of technical alignment, if we're talking about all work that is deemed 'technical alignment.' e.g., I'm not sure that on the margin I would prefer an additional alignment researcher (without knowing what they were researching or anything else about them), though I think it's very unlikely that they would have net-negative impact. So, I think I disagree that (a) "buying time" (excluding weird pivotal acts like trying to shut down labs) might have net negative impact and that & thus also that (b) "buying time" has more variance than technical alignment. edit: Thought about it more and I disagree with my original formulation of the disagreement. I think "buying time" is more likely to be net negative than alignment research, but also that alignment research is usually not very helpful.
9
RobertM
1y
It's not clear to me that the variance of "being a technical researcher" is actually lower than "being a social coordinator". ย Historically, quite a lot of capabilities advancements have come out of efforts that were initially intended to be alignment-focused.ย  Edited to add: I do think it's probably harder to have a justified inside-view model of whether one's efforts are directionally positive or negative when attempting to "buy time", as opposed to "doing technical research", if one actually makes a real effort in both cases.

Out of interest, were you considering students working together and thus submitting similar work as being plagiarism? Or was it more just a lot of cases of some students fully copy/pasting another's work against their wishes?

Sounds like it was a very successful program!

I think this should be acommpanied by a message/prompt in the comment text field that tells people that this post was a draft and to be err in favour of not giving any negative feedback

Sorry for a negative comment, but I think that all of these interventions fail to really address wild animal suffering, and that that is pretty clear already. This is simply due to the fact that pretty much all interventions on WAW have only a temporary positive effect, or worse are zeroed out completely, by the malthusian trap.

  • wind farm deaths: probably not worse than death by predation or starvation. Also other people's hobby horse already.
  • culling: not worse than predation or starvation. IMO often net positive, keeps populations in growth mode. (Of co
... (read more)

Thanks for engaging with the report. I'll offer a response since Tapinder's summer fellowship has ended and I was her manager during the project. I've made a general comment in response to Tristan that applies here too.

On your comment specifically, the "malthusian trap" is empirically not always supported. A population can approach or be at its carrying capacity and still have adequate resources, for instance if they simply do not reproduce as much due to less resource surplus.

Wow people really downvoted it. I just ignored it, in general I don't like to downvote people who are talking about their poor mental health ๐Ÿคทโ€โ™‚๏ธ

1
Rebecca
1y
I think people downvoted it because it comes across more as criticism of other people, rather than the purpose being to talk about the commenter's own mental health

Not sure why people were using the main downvote button on this one, and not just the disagree downvote.

2
Emrik
1y
I think people thought it was bad faith or deliberately provocative or something. Or people aren't used to separating between "disagreement" and "I want to discourage this behaviour". Or maybe they downvoted it because I stated something potentially harmfwl without justifying it, and they were worried that people would defer to me. I think I shouldn't have posted the comment without more explanation, because it predictably would be misunderstood, wouldn't help push communication norms further towards what I want, and made the forum feel less nice for people.
4
Emrik
2y
I made a bad comment, but I tried to explain it in response to Bruce's comment above. Just letting you know, not requesting that you to read it.

I have been attending a Secular Buddhist  group for a couple of years and I have also seen this similarity.

My main idea about how to link EA and Buddhism is as follows:

  1. People often become involved in Buddhism for the meditation and mindfulness - to learn about their own minds, change their experience of the world, and have a better life.
  2. There are a lot of places where having greater impact as an EA is helped by developing positive motivation and different desires, through practicing meditation and mindfulness. 
    1. Eg. For example, the idea of "purcha
... (read more)

No idea how to go about finding information on this, but by my personal priors I would weight various kinds of evidence as follows:

  • ~0 weight on any anecdotal evidence,
  • low weight on studies and clinical trials,
  • moderate weight on arguments by biological plausibility.

Being related to diet, my prior is that people are usually over thinking it. However I have always agreed that it seems unlikely that a fully vegan diet has no nutritional downsides without supplementation.

I've done a cursory search, just wikipedia, here are my thoughts on the biological plausibi... (read more)

7
mako yass
2y
You should only put approximately zero weight on anecdotes that got to you through a sensationalism-maximizing curation system with biases you don't understand, which I hope this wasn't? Regardless, the anecdotes are mostly just meant to be clarifying examples of the kind of thing effect that I am trying to ask about, I don't expect people to pass them along or anything. I decided not to talk about biological plausibility, because I don't get the impression pharmacology or nutrition follows reductive enough rules that anyone can reason a-priori about it very well. It will surprise us, it will do things that don't make sense. I actually wrote out a list of evolutionary stories that end up producing this effect, some of them are quite interesting, but I decided not to post them, because it seemed like ย a distraction. I guess I'll post some theories now though: - This sort of phantasic creativity was not useful in pre-industrial societies, because there was no way to go far beyond the social consensus reality and prove that you were right and do anything with it (that's only the case today because of, basically, the investment cycle, and science and technology, which took hundreds of years to start functioning after it was initially proposed). The body needed an excuse to waste creatine, so in sapiens, it only did it when we ate an abnormal amount of meat, but sudden gluts of meat would occur frequently enough for the adaptation to be maintained. - Or maybe eating lots of meat/fish was kind of the norm for millions of years for dominant populations (I can cite this if you're that interested). And maybe there's a limit to how fast the body can replenish brain-creatine (investigate this assumption, Gregger seemed implicitly sus about it). In that case, we might have an effect where the brain implements creatine frugality by lowering our motivation to think in phosphocreatine-burning ways, which then may lead to a glorification of that frugality, which then becomes stick
2
mako yass
2y
I think it would have gotten better answers there, but I'd guess that if this health trap is real, LWers are much less likely to fall into it (they'd notice the loss), and veganism isn't really as on-topic there.

A defense of the inner ring, excerpts from the original.

...

I must now make a distinction. I am not going to say that the existence of Inner Rings is an Evil. It is certainly unavoidable. There must be confidential discussions: and it is not only a bad thing, it is (in itself) a good thing, that personal friendship should grow up between those who work together. And it is perhaps impossible that the official hierarchy of any organisation should coincide with its actual workings. If the wisest and most energetic people held the highest spots, it might coinci

... (read more)

Hey, appreciate your response. Perhaps we should discuss the meaning of the word "hub" here? To me, it is about 1) Having enough EAs to establish beneficial network effects, and 2) to have a reason why the EAs living there aren't constantly incentivised to move elsewhere (which also means they can live and and work there if they choose)

I think that your value proposition of a beautiful, cheap location for remote work is a great reason for a hub! This fulfills condition 2). Then, having enough people fulfills 1).

However, network effects cause increasing ret... (read more)

1
Tereza_Flidrova
2y
Hey Max, thanks a lot for the response! I don't think I generally disagree with any of the points you are mentioning above. It is important to stress we are not proposing people move to Sandusky, especially at this stage it is about providing a place to run retreats, come and work remotely for a few weeks/months, run fellowship, focus on getting some deep work done while being able to take part in what Sandusky has to offer. This means people from abroad might still be able to come and take part (I, for example, don't have US citizenship). Long term, I would love to work on projects to establish hubs outside of the US (being Czech, I see the benefits of having a thriving EA community outside of the US!). This project, however, is a great way to learn about setting up such places, knowledge I will be able to use when working on future projects. For me, this project is as much about learning as it is about Sandusky as a place in particular, which provides a great platform to test ideas. We are hoping to document the process to create a 'hub-creation package' of some sort to help make such projects run more smoothly every time a new one happens. I hope this explains my thinking a little bit, your comments definitely got me thinking, thanks a lot for bringing them up! Would be happy to discuss this further on a call if you thought that might be helpful.ย 

Hi!

I have to say I strongly disagree with this idea, for one particular reason. If we successfully establish a new hub with cheap living costs and beautiful nature, it MUST be outside the USA. The USA is notoriously hard to immigrate into from most countries!

It is unfortunate that we already have one hub (SF Bay Area / Berkeley) in the USA, although I definitely am OK with D.C. becoming a hub. However, I'd ask any Americans who want to be in an EA hub, but don't want to be in those two places, to go to someone else's hub (Mexico City, Cape Town), or if still wanting to set one up, to do so in a jurisdiction with permissive immigration.

2
Tereza_Flidrova
2y
Hi Max, thanks a lot for this comment!ย  I generally agree that there need to be hubs located outside of the US, and that possibly, this needs to be prioritized. However, I strongly disagree with the premise that we need to choose any single location for an EA hub and reject all others. I think that working on multiple types of hubs in different locations is the way to go. If there are communities within the US who might want to support an EA hub that's easily accessible, yet outside of the core EA hub locations, I fully support this. And if you aren't a US citizen, you can come and spend a few months in the US without a work permit, I don't think my post suggested we want people to move there forever or to immigrate to the US in order to come and utilize the space.ย  With that said, I am all for supporting efforts to create hubs outside of the US though and appreciate you stressing this out.

Yeah the example above with choosing to not get promoted or not recieve funding is a more realistic scenario.

I agree these situations are somewhat rare in practice.

Re. AI Safety, my point was that these situations are especially rare there (among people who agree it's a problem, which is about states of knowledge anyway, not about goals)

Thanks for this post, I think it's a good discussion.

Epistemic status: 2am ramble.

It's about trust, although it definitely varies in importance from situation to situation. There's a very strong trust between people who have strong shared knowledge that they are all utilitarian. Establishing that is where the "purity tests" get value.

Here's a little example.

Let's say you had some private information about a problem/solution that the ea community hadn't yet worked on, and the following choice: A) reveal it to the community, with near certainty that the problem will be solved at least as well as if you yoursel... (read more)

I agree that high-trust networks are valuable (and therefore important to build or preserve). However, I think that trustworthiness is quite disconnected to how people think of their life goals (whether they're utilitarian/altruistic or self-oriented). Instead, I think the way to build high-trust networks is by getting to know people well and paying attention to the specifics.

For instance, we can envision"selfish" people who are nice to others but utilitarians who want to sabotage others over TAI timeline disagreements or disagreements about population eth... (read more)

4
๐•ฎ๐–Ž๐–“๐–Š๐–—๐–†
2y
Huh. If I had a bright idea for AI Safety, I'd share it and expect to get status/credit for doing so. The idea of hiding any bright alignment research ideas I came up with didn't occur to me. I'm under the impression that because of common sense morals (i.e. I wouldn't deliberately sabotage to get the chance to play hero), selfishly motivated EAs like me don't behave particularly different in common scenarios. There are scenarios where my selfishness will be highlighted, but they're very, very narrow states and unlikely to materialise in the real world (highly contrived and only in thought experiment land). In the real world, I don't expect it to be relevant. Ditto for concerns about superrational behaviour. The kind of superrational coordination that's possible for purely motivated EAs but isn't possible with me is behaviour I don't expect to actually manifest in the real world.

Agree with all that yep, and perhaps I should phrase my comment better.

This doesn't address the elephant which is "quality" of talent. EA has a funding overhang with respect to some implicit "quality line" at which people will be hired. Getting more people who can demonstrate talent over that line (where the placement of each specific line is very dependent on context) lowers the funding overhang, but only getting more people under the line doesn't change anything.

5
Owen Cotton-Barratt
2y
Right. (Which is significantly about level of context not just innate properties of the people; also as I alluded to briefly depends on the coordination problem of finding the right roles for people.) But I don't really think introducing this changes the concept of funding overhangs?

No no, I still believe it's a great idea. It just needs people to want to do it, and I was just sharing my observation that there doesn't seem to be that many people who want it enough to offset other things in their life (everyone is always busy).

Your comment about "selecting for people who don't find it boring" is a good re-framing, I like it.

1
Holly Morgan
2y
Oh yes I know - with my reply I was (confusingly) addressing the unreceptive people more than I was addressing you. I'm glad that you're keen :-)

I've had quite a few people ask me "What's altruism?" when running university clubs fair stalls for EA Wellington.

1
Aman Patel
2y
Yeah, I've had several (non-exchange) students ask me what altruism means--my go-to answer is "selflessly helping others," which I hope makes it clear that it describes a practice rather than a dogma.ย 
1
Rebecca
2y
We had that as well with EA USyd, but they were all security guards etc working on the campus, or some exchange students.

I've been very keen to run "deep dives" where we do independent research on some topic, with the aim that the group as a whole ends up with significantly more expertise than at the start.

I've proposed doing this with my group, but people are disappointingly unreceptive to it, mainly because of the time commitment and "boringness".

1
Holly Morgan
2y
Maybe you want to select for the kind of people who don't find it too boring! My guess, though, is that the project idea as currently stated is actually a bit too boring for even most of the people that you'd be trying to reach. And I guess groups aren't keen to throw money at trying to make it more fun/prestigious in the current climate... I've updated away from thinking this is a good idea a little bit, but would still be keen to see several groups try it.

For an overview of most of the current efforts into "epistemic infrastructure", see the comments on my recent post here https://forum.effectivealtruism.org/posts/qFPQYM4dfRnE8Cwfx/project-a-web-platform-for-crowdsourcing-impact-estimates-of

For an overview of most of the current efforts into "epistemic infrastructure", see the comments on my recent post here https://forum.effectivealtruism.org/posts/qFPQYM4dfRnE8Cwfx/project-a-web-platform-for-crowdsourcing-impact-estimates-of

Post civilizational collapse you might not be able to pay that cost though

Buying coal mines to secure energy production post-global-catastrophe is a much more interesting question.

Seems to me that buying coal, rather than mines, is a better idea in that case.

3[anonymous]2y
that's one option, but it would also be more expensive because then you have to cover all the costs of mining.ย 

I'm really hoping we can get some better data on resource allocation and estimated effectiveness to make it clearer when funders or individuals should return to focusing on global poverty etc.

There's a few projects in the works for "ea epistemic infrastructure"

2
Harrison Gietz
2y
Do you have any more info on these "epistemic infrastructure" projects or the people working on them? I would be super curious to look into this more.

Ok - this is a good critique of my comment.

I was kind of off-topic and responding to something a bit more general. Since writing my comment I have found someone on the forum summarizing my perspective better.

https://forum.effectivealtruism.org/posts/bsTXHJFu3Srurbg7K/leftism-virtue-cafe-s-shortform?commentId=mdhfHBe3k5wvqXfo2

 

and relatedly re. funding

https://forum.effectivealtruism.org/posts/bsTXHJFu3Srurbg7K/leftism-virtue-cafe-s-shortform?commentId=eW8zdL2MiXsgNPgMa

Load more