All of peterhartree's Comments + Replies

Hi Jaap—if you'd like to give feedback on narrations, best option is to email me directly. If you'd like to contact the EA Forum team for other reasons, see the contact page.

I can imagine "co-eff" being sticky.

I agree that "co-give" is nice but it's sufficiently harder to say that I predict it won't catch on.

2
NickLaing
yeah co-eff is nice 
6
NickLaing
co-give is nicer?

Congratulations!

The name has six syllables—is there an official line on what the conversational abbreviation should be?

5
Sergi Lange-Soler
Looks like they're using just "Coefficient"
4
peterhartree
I'm defaulting to co-egg...

I've no experience writing questions for prediction markets. With that caveat: something like that question sounds good.

Ideally I'd like to see the 1-year analysis run in 2026Q1.

Notably, in that video, Garry is quite careful and deliberate with his phrasing. It doesn't come across as a case of him doing excited offhand hype. Paul Buchheit nods as he makes the claim.

Cool, thanks. With that source, I agree it's correct to say that Garry Tan has claimed that "YC batches are the fastest growing in their history because of generative AI" for the summer 2024, autumn 2024 and winter 2025 batches.

Have you noticed him making a similar claim for earlier batches?

2
Ben_West🔸
Not to my knowledge. I agree with your point that maybe it's just too soon to see any results (this is why I put "yet" in the title). I've been trying to think about what a good prediction market for this post would be, because I'd like to get some signal on whether this is the explanation - any thoughts? Maybe something like "if I run this exact same analysis one year from now, will at least two 2024 batch companies be on the list of 20 fastest 2-year growth YC startups?"

Thanks for the post. Of your caveats, I'd guess 4(d) is the most important:

Generative AI is very rapidly progressing. It seems plausible that technologies good enough to move the needle on company valuations were only developed, say, six months ago, in which case it would be too early to see any results.

Personally, things have felt very different since o3 (April) and, for coding, the Claude 4 series (May).

Anthropic's run-rate revenue went from $1B in January to $5B in August.

This post misquotes Garry Tan. You wrote (my emphasis):

Y Combinator CEO Garry Tan has publicly stated that recent YC batches are the fastest growing in their history because of generative AI.

But Garry's claim was only about the winter 2025 batch. From the passage you cited:

The winter 2025 batch of YC companies in aggregate grew 10% per week, he said.

“It’s not just the number one or two companies -- the whole batch is growing 10% week on week”

3
Ben_West🔸
He's been saying the same thing for a while. E.g. here: I will add this as a footnote to clarify though, thanks for pointing it out!

Thanks for reporting this.

I've just fixed the narration.

2
calebp
This is great, thanks for sharing. I hadn’t come across it, but it matches my intuitions well.

A long overdue thank you for this comment.

I looked into this, and there is in fact some evidence that less expressive voices are easier to understand at high speed. This factor influenced our decision to stick with Ryan for now.

The results of the listener survey were equivocal. Listener preferences varied widely, with Ryan (our existing voice) and Echo coming out tied and slightly on top, but not with statistical significance. [1]

Given that, we plan to stick with Ryan for now. Two considerations that influence this call, independently of the survey result:

  1. There's an habituation effect, such that switching costs for existing hardcore listeners are significant.
  2. There's some evidence that more expressive voices are less comprehensible at high listening speeds. Ryan is less expressive
... (read more)

Similar reasons that I want both of Twitter's "home" and "following" views.

I want "People you follow" to include everything and it should be chronological.

If quick takes are includedl, I guess the "People you follow" content could dominate the items that currently make the "New and upvoted" section.

Quick thought is that I'd want a separate section on the homepage called "People you follow".

Mixing it into the current "New & Upvoted" section seems bad to me but that's really just a hot take.

2
Toby Tremlett🔹
V reasonable- do you have an idea of why it seems bad?   

I wish it was easier to find the best social media discussions of forum posts and quick takes.

E.g. Ozzie flagged some good Facebook discussion of a recent quick take [1], but people rarely do that.

Brainstorm on minimal things you could do:

a. Link to Twitter search results (example). (Ideally you might show the best Twitter discussions in the Forum UI, but I'm not sure if their API terms would permit that.)

b. Encourage authors to link to discussion on other platforms, and have a special place in the UI that makes such links easy to find.


  1. I actually can't

... (read more)

Hmm I think Twitter-ish feed containing posts, quick takes and comments from just the users I follow would be very useful. Not sure what it'd do to the forum dynamic overall though, but could be good, actually?

4
Toby Tremlett🔹
Just brainstorming here but what if quick takes from users you follow were interleaved into your forum frontpage feed? I.e. they would appear similarly to frontpage posts posts, but probably with an icon, and then expand on click. 

Another way to make me engage more with "Quick takes" is if I could get a feed of takes from users I follow. I can kinda get that in the "Notifications" list but again it's super hard to skim the list to find the ones of interest.

2
peterhartree
Hmm I think Twitter-ish feed containing posts, quick takes and comments from just the users I follow would be very useful. Not sure what it'd do to the forum dynamic overall though, but could be good, actually?

I ~never check the "Quick takes" section even though I bet it has some great stuff.

The reason is that its too hard to find takes that are on a topic I care about, and/or have an angle that seems interesting.

I think I'd be much more likely to read quick takes if the takes had titles, and/or if I could filter them by topic tags.

Maybe you can LLM-generate default title and tag suggestions for the author when they're about to post?

2
peterhartree
Another way to make me engage more with "Quick takes" is if I could get a feed of takes from users I follow. I can kinda get that in the "Notifications" list but again it's super hard to skim the list to find the ones of interest.

I wish the notifications icon at the top right would have have separate icon to indicate replies to my posts and my comments.

At the moment those get drowned out by notifications about posts and comments from people I follow (which are so numerous that I've been trained to rarely check the notification inbox, because it's such a rabbit hole...), so I often see comments weeks or months too late.

I'd also like to be able to get email notifications for just replies to my posts and replies to my comments.

2
Sarah Cheng 🔸
Thanks for the suggestion! I like the idea of visually distinguishing the most important notifications (i.e. the ones you might want to reply to). There are a lot of customization options for notifications, so I'd recommend looking through your account settings: For example, you can set "Comments on posts/events I'm subscribed to" and "Replies to my comments" to notify you both on-site and via email. You may also prefer to batch email notifications (rather than receive them immediately) if you get too many (for example, you could batch "Comments by users I'm subscribed to" to get an email about it once per day).

We made this change a few weeks ago. I'm sorry for the delay—I didn't see your message until now. I've tweaked my notification setup so that I'll see messages on this thread sooner.

1
EffectiveAdvocate🔸
Thank you for fixing this! 

Thanks. We're now filtering the diamond emojis out of the narrations. I've left the others in for now.

1
Lovkush 🔸
I found that out today! Need to update my recommendation.

1. My current process

I check a couple of sources most days, at random times during the afternoon or evening. I usually do this on my phone, during breaks or when I'm otherwise AFK. My phone and laptop are configured to block most of these sources during the morning (LeechBlock and AppBlock).

When I find something I want to engage with at length, I usually put it into my "Reading inbox" note in Obsidian, or into my weekly todo list if it's above the bar.

I check my reading inbox on evenings and weekends, and also during "open" blocks that I sometimes schedule... (read more)

Thanks for your feedback.

For now, we think our current voice model (provided by Azure) is the best available option all things considered. There are important considerations in addition to human-like delivery (e.g. cost, speed, reliability, fine-grained control).

I'm quite surprised that an overall-much-better option hasn't emerged before now. My guess is that something will show up later in 2024. When it does, we will migrate.

2
peterhartree
Update: we're seeking feedback on some new voice models.

There are good email newsletters that aren't reliably read.

Readit.bot turns any newsletter into a personal podcast feed.

TYPE III AUDIO works with authors and orgs to make podcast feeds of their newsletters—currently Zvi, CAIS, ChinAI and FLI EU AI Act, but we may do a bunch more soon.

I think that "awareness of important simple facts" is a surprisingly big problem.

Over the years, I've had many experiences of "wow, I would have expected person X to know about important fact Y, but they didn't".

The issue came to mind again last week:

My sense is that many people, including very influential folks, could systematically—and efficiently—improve their awareness of "simple important facts".

There may be quick wins here. For example, there are existing tools that aren't widely used (e.g. Twitter lists; Tweetdeck). There are good email newsletters ... (read more)

4
peterhartree
Readit.bot turns any newsletter into a personal podcast feed. TYPE III AUDIO works with authors and orgs to make podcast feeds of their newsletters—currently Zvi, CAIS, ChinAI and FLI EU AI Act, but we may do a bunch more soon.

Bret Taylor and Larry Summers (members of the current OpenAI board) have responded to Helen Toner and Tasha McCauley in The Economist.

The key passages:

Helen Toner and Tasha McCauley, who left the board of Openai after its decision to reverse course on replacing Sam Altman, the CEO, last November, have offered comments on the regulation of artificial intelligence (AI) and events at OpenAI in a By Invitation piece in The Economist.

We do not accept the claims made by Ms Toner and Ms McCauley regarding events at OpenAI. Upon being asked by the former board (

... (read more)

Andrew Mayne points out that “the base model for ChatGPT (GPT 3.5) had been publicly available via the API since March 2022”.

On (1): it's very unclear how ownership could be compatible with no financial interest.

Maaaaaybe (2) explains it. That is: while ownership does legally entail financial interest, it was agreed that this was only a pragmatic stopgap measure, such that in practice Sam had no financial interest.

Thanks for the context!

Obvious flag that this still seems very sketchy. "the easiest way to do that due to our structure was to put it in Sam's name"? Given all the red flags that this drew, both publicly and within the board, it seems hard for me to believe that this was done solely "to make things go quickly and smoothly."

I remember Sam Bankman-Fried used a similar argument around registering Alameda - in that case, I believe it later led to him later having a lot more power because of it.
 

For context:

  1. OpenAI claims that while Sam owned the OpenAI Startup Fund, there was “no personal investment or financial interest from Sam”.
  2. In February 2024, OpenAI said: “We wanted to get started quickly and the easiest way to do that due to our structure was to put it in Sam's name. We have always intended for this to be temporary.”
  3. In April 2024 it was announced that Sam no longer owns the fund.
5
peterhartree
On (1): it's very unclear how ownership could be compatible with no financial interest. Maaaaaybe (2) explains it. That is: while ownership does legally entail financial interest, it was agreed that this was only a pragmatic stopgap measure, such that in practice Sam had no financial interest.

Sam didn't inform the board that he owned the OpenAI Startup Fund, even though he constantly was claiming to be an independent board member with no financial interest in the company.

Sam has publicly said he has no equity in OpenAI. I've not been able to find public quotes where Sam says he has no financial interest in OpenAI (does anyone have a link?).

1
James Payor
He said this during that initial Senate hearing iirc, and I think he was saying this line frequently around then (I recall a few other instances but don't remember where).
3
Benevolent_Rain
It would be hard to imagine he has no interest, I would say even a simple bonus scheme whether stock, options, cash, etc. would count as "interest". If company makes money then so does he.

From the interview:

When ChatGPT came out, November 2022, the board was not informed in advance about that. We learned about ChatGPT on Twitter.

Several sources have suggested that the ChatGPT release was not expected to be a big deal. Internally, ChatGPT was framed as a “low-key research preview”. From The Atlantic:

The company pressed forward and launched ChatGPT on November 30. It was such a low-key event that many employees who weren’t directly involved, including those in safety functions, didn’t even realize it had happened. Some of those who were aware

... (read more)

Andrew Mayne points out that “the base model for ChatGPT (GPT 3.5) had been publicly available via the API since March 2022”.

I think what would be more helpful for me is the other things discussed in board meetings. Even if GPT was not expected to be a big deal, if they were (hyperbolic example) for example discussing whether to have a coffee machine at the office, I think not mentioning ChatGPT would be striking. On the other hand, if they only met once a year and only discussed e.g. if they are financially viable or not, then perhaps not mentioning ChatGPT makes more sense. And maybe even this is not enough - it would also be concerning if some board members wanted more info, ... (read more)

You wrote:

[OpenAI do] very little public discussion of concrete/specific large-scale risks of their products and the corresponding risk-mitigation efforts (outside of things like short-term malicious use by bad API actors, where they are doing better work).

This doesn't match my impression.

For example, Altman signed the CAIS AI Safety Statement, which reads:

Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.

The “Preparedness” page—linked from the top navigation... (read more)

3
Ozzie Gooen
I agree that there's a lot of evidence that people at OpenAI have thought that AI could be a major risk, and I think that these are good examples. I said here, "concrete/specific large-scale risks of their products and the corresponding risk-mitigation efforts (outside of things like short-term malicious use by bad API actors, where they are doing better work)." Just looking at the examples you posted, most feel pretty high-level and vague, and not very related to their specific products.  > For example, Altman signed the CAIS AI Safety Statement, which reads... This was a one-sentence statement. It easily sounds to me like saying, "Someone should deal with this, but not exactly us." > The framework itself goes into more detail, proposing scorecards for assessing risk in each category. I think this is a good step, but it seems pretty vague to me. There's fairly little quantifiable content here, a lot of words like "medium risk" and "high risk".  From what I can tell, the "teeth" in the document is, "changes get brought up to management, and our board", which doesn't fill me with confidence.  Related, I'd be quite surprised if they actually followed through with this much in the next 1-3 years, but I'd be happy to be wrong! 

OpenAI's “Planning for AGI & Beyond” blog post includes the following:

As our systems get closer to AGI, we are becoming increasingly cautious with the creation and deployment of our models. Our decisions will require much more caution than society usually applies to new technologies, and more caution than many users would like. Some people in the AI field think the risks of AGI (and successor systems) are fictitious; we would be delighted if they turn out to be right, but we are going to operate as if these risks are existential.

At some point, the ba

... (read more)

I have bad feelings about a lot of this.

The "Planning for AGI & Beyond" doc seems to me to be heavily inspired by a few other people at OpenAI at the time, mainly the safety team, and I'm nervous those people have less influence now. 

At the bottom, it says:

Thanks to Brian Chesky, Paul Christiano, Jack Clark, Holden Karnofsky, Tasha McCauley, Nate Soares, Kevin Scott, Brad Smith, Helen Toner, Allan Dafoe, and the OpenAI team for reviewing drafts of this.

Since then, Tasha and Helen have been fired off the board, and I'm guessing relations have soured with others listed. 

Yes. (4) and (11) are also very much "citation needed". My sense is that they would need to be significantly moderated to fit the facts (e.g. the profit cap is still a thing).

On short stories, two notable examples: Narrative Ark by Richard Ngo and the regular "Tech Tales" section at the end of Jack Clark's Import AI newsletter.

I read the paper, then asked Claude 3 to summarise. I endorse the following summary as accurate:

The key argument in this paper is that Buddhism is not necessarily committed to pessimistic views about the value of unawakened human lives. Specifically, the author argues against two possible pessimistic Buddhist positions:

  1. The Pessimistic Assumption - the view that any mental state characterized by dukkha (dissatisfaction, unease, suffering) is on balance bad. 
  2. The Pessimistic Conclusion - the view that over the course of an unawakened life, dukkha will al
... (read more)
1[anonymous]
Wow, this is good - go Claude 3! 

Thanks for this. 

I take the central claim to be:

Even if an experience contains some element of dukkha, it can still be good overall if its positive features outweigh the negative ones. The mere presence of dukkha does not make an experience bad on balance.

I agree with this, and also agree that it's often overlooked.

The completion rate at BlueDot Impact averaged out at about 75%

How do you define completion?

1
Jamie B
What I had in mind was "shows up to all 8 discussion groups for the taught part of the course". I also didn't check this figure, so that was from memory. True, there are lots of ways to define it (e.g. finishing the readings, completing the project, etc)

I think so. I'll put a note about this at the top of the post.

Perhaps it’s just the case that the process of moral reflection tends to cause convergence among minds from a range of starting points, via something like social logic plus shared evolutionary underpinnings.

Yes. And there are many cases where evolution has indeed converged on solutions to other problems[1].

  1. ^

    Some examples:

    (Copy-pasted from Claude 3 Opus. They pass my eyeball fact-check.)

    1. Wings: Birds, bats, and insects have all independently evolved wings for flight, despite having very different ancestry.
    2. Eyes: Complex camera-like eyes have evolved inde
... (read more)

My own attraction to a bucket approach (in the sense of (1) above) is motivated by a combination of:

(a) reject the demand for commensurability across buckets.

(b) make a bet on plausible deontic constraints e.g. duty to prioritise members of the community of which you are a part.

(c) avoid impractical zig-zagging when best guess assumptions change.

Insofar as I'm more into philosophical pragmatism than foundationalism, I'm more inclined to see a messy collection of reasons like these as philosophically adequate.

I think there are two things to justify here:

  1. The commitment to a GHW bucket, where that commitment involves "we want to allocate roughly X% of our resources to this".

  2. The particular interventions we fund within the GHW resource bucket.

I think the justification for (1) is going to look very different to the justification for (2).

I'm not sure which one you're addressing, it sounds like more (2) than (1).

2
Richard Y Chappell🔸
I'm more interested in (1), but how we justify that could have implications for (2).
2
peterhartree
My own attraction to a bucket approach (in the sense of (1) above) is motivated by a combination of: (a) reject the demand for commensurability across buckets. (b) make a bet on plausible deontic constraints e.g. duty to prioritise members of the community of which you are a part. (c) avoid impractical zig-zagging when best guess assumptions change. Insofar as I'm more into philosophical pragmatism than foundationalism, I'm more inclined to see a messy collection of reasons like these as philosophically adequate.

Would you be up for spelling out the problem of "lacks adequate philosophical foundations"?

What criteria need to be satisfied for the foundations to be adequate, to your mind?

Do they e.g. include consequentialism and a strong form of impartiality?

3
peterhartree
I think there are two things to justify here: 1. The commitment to a GHW bucket, where that commitment involves "we want to allocate roughly X% of our resources to this". 2. The particular interventions we fund within the GHW resource bucket. I think the justification for (1) is going to look very different to the justification for (2). I'm not sure which one you're addressing, it sounds like more (2) than (1).

I hesitate to post things like this, because “short, practical advice” posts aren't something I often see on the Forum.

I'm not sure if this is the kind of thing that's worth encouraging as a top-level post.

In general I would like to read more posts like this from EA Forum users, but perhaps not as part of the front page.

3
peterhartree
Some examples of the kinds of thing I might share, were there an obvious place to do so: * How to start a blog in 5 seconds for $0 * Unfortunately, PDF copies of many academic books & scientific papers can be easily found on the Anna's Archive website * Use Airalo to get mobile data in any country in less than 5 minutes

Thanks for this. I'd be keen to see a longer list of the interesting for-profits in this space.

Biobot Analytics (wastewater monitoring) are the only for-profit on the 80,000 Hours job board list.

3
Alex D 🔸
This is a pretty good overview: https://www.decodingbio.com/p/decoding-biosecurity-and-biodefense I know the space reasonably well, happy to connect and discuss with anyone interested!
Load more