All of Lizka's Comments + Replies

I'm also worried about an "epistemics" transformation going poorly, and agree that how it goes isn't just a question of getting the right ~"application shape" — something like differential access/adoption[1] matters here, too.

@Owen Cotton-Barratt, @Oliver Sourbut, @rosehadshar and I have been thinking a bit about these kinds of questions, but not as much as I'd like (there's just not enough time). So I'd love to see more serious work on things like "what might it look for our society to end up with much better/worse epistemic infrastructure (and how m... (read more)

3
Oliver Sourbut
Basically +1 here. I guess some relevant considerations are the extent to which a tool can act as antidote to its own (or related) misuse - and under what conditions of effort, attention, compute, etc. If that can be arranged, then 'simply' making sure that access is somewhat distributed is a help. On the other hand, it's conceivable that compute advantages or structural advantages could make misuse of a given tech harder to block, in which case we'd want to know that (without, perhaps, broadcasting it indiscriminately) and develop responses. Plausibly those dynamics might change nonlinearly with the introduction of epistemic/coordination tech of other kinds at different times. In theory, it's often cheaper and easier to verify the properties of a proposal ('does it concentrate power?') than to generate one satisfying given properties, which gives an advantage to a defender if proposals and activity are mostly visible. But subtlety and obfuscation and misdirection can mean that knowing what properties to check for is itself a difficult task, tilting the other way. Likewise, narrowly facilitating coordination might produce novel collusion with substantial negative externalities on outsiders. But then ex hypothesi those outsiders have an outsized incentive to block that collusion, if only they can foresee it and coordinate in turn. It's confusing.

I didn't end up writing a reflection in the comments as I'd meant to when I posted this, but I did end up making two small paintings inspired by Benjamin Lay & his work. I've now shared them here

I think of today (February 8) as "Benjamin Lay Day", for what it's worth. (Funny timing :) .) 

Another one I'd personally add might be November 4 for Joseph Rotblat. And just in case you haven't seen / just for reference, there are some related resources on the Forum, e.g. here https://forum.effectivealtruism.org/topics/events-on-the-ea-forum, and here https://forum.effectivealtruism.org/posts/QFfWmPPEKXrh6gZa3/the-ea-holiday-calendar

In fact I think the Forum team may also still maintain a list/calendar of possible days to celebrate somewhere. ... (read more)

Lizka
58
1
0
20

Benjamin Lay — "Quaker Comet", early (radical) abolitionist, general "moral weirdo" — died on this day 267 years ago. 

I shared a post about him a little while back, and still think of February 8 as "Benjamin Lay Day". 

...

Around the same time I also made two paintings inspired by his life/work, which I figured I'd share now. One is an icon-style-inspired image based on a portrait of him[1]:

Benjamin Lay portrait, in front of his cave (stylized) and the quaker meeting house in Pennsylvania

The second is based on a print depicting the floor plan of an infamous slave ship (Brooks). The print was used by abolitionists (mainly(?) the Society for Effec... (read more)

2
Alex (Αλέξανδρος)
Thank you for reminding about this remarkable person. I'll add him to my personal inspirational list of Humanity's Best People
5
Angelina Li
Wow, I've never seen that print before. That is absolutely horrifying. I feel kind of sick looking at it. What a stark reminder of the costs of getting morality wrong. Thank you for painting it, for sharing it, and for the reminder of this day.

When thinking about the impacts of AI, I’ve found it useful to distinguish between different reasons for why automation in some area might be slow. In brief: 

  1. raw performance issues
  2. trust bottlenecks
  3. intrinsic premiums for “the human factor”
  4. adoption lag
  5. motivated/active protectionism towards humans

I’m posting this mainly because I’ve wanted to link to this a few times now when discussing questions like "how should we update on the shape of AI diffusion based on...?". Not sure how helpful it will be on its own!


In a bit more detail:

(1) Raw performance issue... (read more)

Yeah, I guess I don't want to say that it'd be better if the team had people who are (already) strongly attached to various specific perspectives (like the "AI as a normal technology" worldview --- maybe especially that one?[1]). And I agree that having shared foundations is useful / constantly relitigating foundational issues would be frustrating. I also really do think the points I listed under "who I think would be a good fit" — willingness to try on and ditch conceptual models, high openness without losing track of taste, & flexibility — matter, an... (read more)

4
William_MacAskill
I disagree with this. There would need to be some other reason for why they should work at Forethought rather than elsewhere, but there are plausible answers to that — e.g. they work on space governance, or they want to write up why they think AI won't change the world much and engage with the counterarguments. 
2
Tax Geek
On the "AI as normal technology" perspective - I don't think it involves a strong belief that AI won't change the world much. The authors restate their thesis in a later post: The idea of focusing more on the deployment stage seems pretty consistent with Will MacAskill's latest forum post about making the transition to a post-AGI society go well. There are other aspects of the "AI as normal technology" worldview that I expect will conflict more with Forethought's, but I'm not sure that conflict would necessarily be frustrating and unproductive - as you say, it might depend on the person's characteristics like openness and willingness to update, etc.
2
OscarD🔸
Nice, yes I think we roughly agree! (Though maybe you are nobler than me in terms of finding a broader range of views provocatively plausible and productive to engage with.)

ore Quick sketch of what I mean (and again I think others at Forethought may disagree with me):

  • I think most of the work that gets done at Forethought builds primarily on top of conceptual models that are at least in significant part ~deferring to a fairly narrow cluster of AI worldviews/paradigms (maybe roughly in the direction of what Joe Carlsmith/Buck/Ryan have written about)
    • (To be clear, I think this probably doesn't cover everyone, and even when it does, there's also work that does this more/less, and some explicit poking at these worldviews, etc.)
    • So
... (read more)
8
MaxDalton
Thanks for writing the post and this comment, Lizka!   I agree that most of Forethought (apart from you!) have views that are somewhat similar to Joe/Buck/Ryan's, but I think that's mostly not via deferral? +1 to wanting people who can explore other perspectives, like Gradual Disempowerment, coalitional agency, AI personas, etc. And the stuff that you've been exploring! I also agree that there's some default more welfarist / consequentialist frame, though I think often we don't actually endorse this on reflection. Also agree that there's some shared thinking styles, though I think there's a bit more diversity in training (we have people who majored in history, CS, have done empirical ML work, etc). Also maybe general note, that on many of the axes you're describing you are adding some of the diversity that you want, so Forethought-as-a-whole is a bit more diverse on these axes than Forethought-minus-Lizka.
6
OscarD🔸
I think I agree maybe ~80%. My main reservation (although quite possibly we agree here) is that if Forethought hired e.g. the 'AI as a normal technology' people, or anyone with equivalently different baseline assumptions and ways of thinking to most of Forethought, I think that would be pretty frustrating and unproductive. (That said, I think brining people like that in for a week or so might be great, to drill down into cruxes and download each others' world models more.) I think there is something great about having lots of foundational things in common with the people you work closely with. But I agree that having more people who share some basic prerequisites of thinking ASI is possible, likely to come this century, being somewhat longtermist and cosmopolitan and altruistic, etc, but disagree a lot on particular topics like AI timelines and threat models and research approaches and so forth can be pretty useful.

Ah, @Gregory Lewis🔸  says some of the above better. Quoting his comment

  • [...]
  • So ~everything is ultimately an S-curve. Yet although 'this trend will start capping out somewhere' is a very safe bet, 'calling the inflection point' before you've passed it is known to be extremely hard. Sigmoid curves in their early days are essentially indistinguishable from exponential ones, and the extra parameter which ~guarantees they can better (over?)fit the points on the graph than a simple exponential give very unstable estimates of the putative ceiling the
... (read more)

I tried to clarify things a bit in this reply to titotal: https://forum.effectivealtruism.org/posts/iJSYZJJrLMigJsBeK/lizka-s-shortform?commentId=uewYatQz4dxJPXPiv 

In particular, I'm not trying to make a strong claim about exponentials specifically, or that things will line up perfectly, etc. 

(Fwiw, though, it does seem possible that if we zoom out, recent/near-term population growth slow-downs might be functionally a ~blip if humanity or something like it leaves the Earth. Although at some point you'd still hit physical limits.)

Oh, apologies: I'm not actually trying to claim that things will be <<exactly.. exponential>>. We should expect some amount of ~variation in progress/growth (these are rough models, we shouldn't be too confident about how things will go, etc.), what's actually going on is (probably a lot) more complicated than a simple/neat progression of new s-curves, etc.

The thing I'm trying to say is more like: 

  • When we've observed some datapoints about a thing we care about, and they seem to fit some overall curve (e.g. exponential growth) reasonably we
... (read more)
4
Eli Rose🔸
"and has performed well" seems like a good crux to zoom in on; for which reference class of empirical trends is this true, and how true is it? It's hard to disagree with "place some weight"; imo it always makes sense to have some prior that past trends will continue. The question is how much weight to place on this heuristic vs. more gears-level reasoning. For a random example, observers in 2009 might have mispredicted Spanish GDP over the next ten years if they placed a lot of weight on this prior.
4
Lizka
Ah, @Gregory Lewis🔸  says some of the above better. Quoting his comment: 

Replying quickly, speaking only for myself[1]:

  1. I agree that the boundary between Community & non-Community posts is (and has always been) fuzzy
    1. You can see some guidance I wrote in early 2023: Community vs. other[2] (a "test-yourself" quiz is also linked in the doc)
      1. Note: I am no longer on the Online Team/an active moderator and don't actually know what the official guidance is today.[3]
    2. I also agree that this fuzziness is not a trivial issue, & it continues to bug me
      1. when I go to the Frontpage, I relatively frequently see posts that I think should
... (read more)
4
Julia_Wise🔸
I agree that separating out community posts was not just a reaction to the FTX situation. Early in CEA's time running the Forum, the community section was an entirely different page, as you can see in this 2019 Wayback capture.  

I sometimes see people say stuff like:

Those forecasts were misguided. If they ended up with good answers, that's accidental; the trends they extrapolated from have hit limits... (Skeptics get Bayes points.)

But IMO it's not a fluke that the "that curve is going up, who knows why" POV has done well. 

A sketch of what I think happens:

Sketch of the discussion about how to extrapolate from curves (see below the image)

There’s a general dynamic here that goes something like:

  1. Some people point to a curve going up (and maybe note some underlying driver)
  2. Others point out that the drivers have inherent constraints (this is an s-curve, not an ex
... (read more)

I'm skeptical of an "exponentials generally continue" prior which is supposed to apply super-generally. For example, here's a graph of world population since 1000 AD; it's an exponential, but actually there are good mechanistic reasons to think it won't continue along this trajectory. Do you think it's very likely to?

The Human Population – Introductory Biology: Evolutionary and ...

8
titotal
And why, exactly, would you expect every single new development to show up at exactly the right time as to make the overall curve remain exponential? What your view actually predicts is that progress will be a series of S-curves... but it says nothing about how long the flat bits in between will be.  Even within the history of AI, we have seen S-curves flatten out: there have been AI winters that lasted literal decades. 

Yeah, this sort of thing is partly why I tend to feel better about BOTECs like (writing very quickly, tbc!):

What could we actually accomplish if we (e.g.) doubled (the total stock/ flow of) investment in ~technical AIS work (specifically the stuff focused on catastrophic risks, in this general worldview)? (you could broaden if you wanted to, obviously)

Well, let's see: 

  • That might look like:
    • adding maybe ~400(??) FTEs similar (in ~aggregate) to the folks working here now, distributed roughly in proportion to current efforts / profiles — plus the funding
... (read more)

Thank you! I used Procreate for these (on an iPad).[1]

(I also love Excalidraw for quick diagrams, have used & liked Whimsical before, and now also semi-grudgingly appreciate Canva.)


Relatedly, I wrote a quick summary of the post in a Twitter thread a few days ago and added two extra sketches there. Posting here too in case anyone finds them useful: 

quick flowchart on how I relate to an ITN estimate
(chaotic) overview of the pitfalls of ITN BOTECs that I discuss
  1. ^

    (And a meme generator for the memes.)

Yeah actually I think @Habryka [Deactivated] discusses these kinds of dynamics here: https://www.lesswrong.com/posts/4NFDwQRhHBB2Ad4ZY/the-filan-cabinet-podcast-with-oliver-habryka-transcript 

Excerpt (bold mine, Habryka speaking):

One of the core things that I was always thinking about with LessWrong, and that was my kind of primary analysis of what went wrong with previous LessWrong revivals, was [kind of] an iterated, [the term] “prisoner's dilemma” is overused, but a bit of an iterated prisoner's dilemma or something where, like, people needed to ha

... (read more)
2
Sarah Cheng 🔸
Yeah I definitely have this in my head when thinking about how to run the EA Forum. But I haven't made a commitment to personally run the site for five years (I'm not a commitment sort of person in general). Maybe that means I'm not a good fit for this role? I also hear conflicting views on whether it's good or bad to "signal that there is real investment". I think I intuitively agree with Habryka here, but then others tell me that it can look bad for us to talk about doing work that doesn't tie directly to impact — like maybe if we talk about improving the UX of the site, people will think that we are wasting charitable money, and that will decrease some people's trust in our team. So for some people, I think they would trust us more if we were doing less work on the site?

Re not necessarily "optimizing" for the Forum, I guess my frame is:

The Online Team is the current custodian of an important shared resource (the Forum). If the team can't actually commit to fulfilling its "Forum custodian" duties, e.g. because the priorities of CEA might change, then it should probably start trying to (responsibly) hand that role off to another person/group. 

(TBC this doesn't mean that Online should be putting all of its efforts into the Forum, just like a parent has no duty to spend all their energy caring for their child. And it's n... (read more)

2
Sarah Cheng 🔸
I agree with this, though I feel like the devil is in the details of what "Forum custodian" means. FWIW I don't think anyone at CEA is interested in shutting down the Forum, or reducing the moderation capacity. Maybe a useful example of "new engineering work" is: we might want to start using the "rejected content" feature that LW has, but we'd need an engineer to update the codebase to enable it on the Forum. So under a strict "no new engineering work" policy, we couldn't start rejecting content, and in fact there's a lot of moderation we couldn't do. We are still doing some engineering work, but we broadly need to justify any work we do under CEA's new strategy. Maybe you think that, if we fail to justify this work under CEA's strategy, but we still think it's valuable to do, then that's the point at which we should start handing the Forum off to someone else?
6
Lizka
Yeah actually I think @Habryka [Deactivated] discusses these kinds of dynamics here: https://www.lesswrong.com/posts/4NFDwQRhHBB2Ad4ZY/the-filan-cabinet-podcast-with-oliver-habryka-transcript  Excerpt (bold mine, Habryka speaking):

I think going for Option 2 ("A bulletin board") or 3 ("Shut down") would be pretty a serious mistake, fwiw. (I have fewer/weaker opinions on 1, although I suspect I'm more pessimistic about it than ~most others.)

...

An internal memo I wrote in early 2023 (during my time on the Online Team) seems relevant, so I've made a public copy: Vision / models / principles for the Forum and the Forum+ team[1]

I probably no longer believe some of what I wrote there, but still endorse the broad points/models, which lead me think, among other things:

  1. That people would proba
... (read more)
6
Sarah Cheng 🔸
Hey Lizka! I love that memo and I agree with most of it (I don't have any particular disagreements, I just feel unsure about some things). It's been a significant influence on the Online Team overall, and on how I think about running the Forum. I also agree with the specific points in your comment. Part of the goal of the exercise was to, as the Online Team, "stare into the abyss" and try to figure out, how much does it really make the world better for us to put capacity towards the Forum? Are we only putting resources towards the Forum because of momentum/personal interest/job security/etc, or do we think that there is actually counterfactual value? Some additional context is that CEA is [moving toward becoming] more of a unified organization now than it has been in the past. My understanding is that we can broadly only do work that aligns with CEA's overall strategy: And, I believe that everyone on the Online Team does want to do the work that is most impactful overall, whether or not that involves the Forum. So part of that equation is, what are the costs (in terms of "impact") of us putting less resources towards the Forum? For example, it's possible that having our product/engineers work on EA Funds would be a more impactful use of their time, and it's also possible that product/engineering work on both projects is valuable enough that we should hire enough people to cover both the EA Forum and EA Funds.
Lizka
49
1
0
17
1

I just want to say that I'm really, really glad that I got the chance to work with JP, and I think JP's work has contributed a bunch for the Forum. 

Below I'm sharing some quick misc notes on why that is. 

Meta: sharing a public comment like this on a friend's post is pretty weird for me! But JP's work has mostly stayed behind the scenes and has thus gotten significantly less acknowledgement than it deserves, IMO, so I did feel the urge to post this kind of thing publicly. As a compromise between those feelings, y'all are getting the comment along ... (read more)

8
Linch
I enjoyed reading this. It's often useful/interesting to build models of what makes someone good at their job, and this particular one is interesting/somewhat surprising to me at parts, even though I know both of you personally and I'm obviously familiar with the forum.
9
JP Addison🔸
Thanks, Lizka. I think this is one of the best retirement presents I could have gotten.

I think this is a good question, and it's something I sort of wanted to look into and then didn't get to! (If you're interested, I might be able to connect you with someone/some folks who might know more, though.)

Quick general takes on what private companies might be able to do to make their tools more useful on this front (please note that I'm pretty out of my depth here, so take this with a decent amount of salt -- and also this isn't meant to be prioritized or exhaustive): 

  • Some of the vetting/authorization processes (e.g. FedRAMP) are burdensome, a
... (read more)

the main question is how high a priority this is, and I am somewhat skeptical it is on the ITN pareto frontier. E.g. I would assume plenty of people care about government efficiency and state capacity generally, and a lot of these interventions are generally about making USG more capable rather than too targeted towards longtermist priorities.

Agree that "how high-priority should this be" is a key question, and I'm definitely not sure it's on the ITN pareto frontier! (Nice phrase, btw.) 

Quick notes on some things that raise the importance for me, thoug... (read more)

I imagine there might be some very clever strategies to get a lot of the benefits of AI without many of the normal costs of integration.

For example:

  1. The federal government makes heavy use of private contractors. These contractors are faster to adopt innovations like AI.
  2. There are clearly some subsets of the government that matter far more than others. And there are some that are much easier to improve than others.
  3. If AI strategy/intelligence is cheap enough, most of the critical work can be paid for by donors. For example, we have a situation where there's a
... (read more)

 Thanks for this comment! I don’t view it as “overly critical.”

Quickly responding (just my POV, not Forethought’s!) to some of what you brought up ---

(This ended up very long, sorry! TLDR: I agree with some of what you wrote, disagree with some of the other stuff / think maybe we're talking past each other. No need to respond to everything here!)

A. Motivation behind writing the piece / target audience/ vibe / etc.

Re:

…it might help me if you explained more about the motivation [behind writing the article] [...] the article reads like you decided the co

... (read more)
9
MichaelDickens
Thanks, this comment gives me a much better sense of where you're coming from. I agree and disagree with various specific points, but I won't get into that since I don't think we will resolve any disagreements without an extended discussion. What I will say is that I found this comment to be much more enlightening than your original post. And whereas I said before that the original article didn't feel like the output of a reasoning process, this comment did feel like that. At least for me personally, I think whatever mental process you used to write this comment is what you should use to write these sorts of articles, because whatever process you used to write this comment, it worked. I don't know what's going on inside your head, but if I were to guess, perhaps you didn't want to write an article in the style of this comment because it's too informal or personal or un-authoritative. Those qualities do make it harder to (say) get a paper published in an academic journal, but I prefer to read articles that have those qualities. If your audience is the EA Forum or similar, then I think you should lean into them. I don't think you were LARPing research, your comment shows a real thought process behind it. After reading all your line item responses, I feel like I understand what you were trying to say. Like on the few parts I quoted as seeming contradictory, I can now see why they weren't actually contradictory and they were part of a coherent stance.

Quick (visual) note on something that seems like a confusion in the current conversation:


Others have noted similar things (eg, and Will’s earlier take on total vs human extinction). You might disagree with the model (curious if so!), but I’m a bit worried that one way or another people are talking past each other (least from skimming the discussion). 

(Commenting via phone, sorry for typos or similar!)

What actually changes about what you’d work on if you concluded that improving the future is more important on the current margin than trying to reduce the chance of (total) extinction (or vice versa)? 

Curious for takes from anyone!

6
OscarD🔸
It felt surprisingly hard to come up with important examples of this, I think because there is some (suspicious?) convergence between both extinction prevention and trajectory changing via improving the caution and wisdom with which we transition to ASI. This both makes extinction less likely (through more focus on alignment and control work, and perhaps slowing capabilities progress or differential accelerating safety-oriented AI applications) and improves the value of surviving futures (by making human takeovers, suffering digital minds etc less likely). But maybe this is just focusing on the wrong resolution. Breaking down 'making the ASI transition wiser', if we are mainly focused on extinction, AI control looks especially promising but less so otherwise. Digital sentience and rights work looks better if trajectory changes dominate, though not entirely. Improving company and government (especially USG) understanding of relevant issues seems good for both. Obviously, asteroids, supervolcanoes, etc work looks worse if preventing extinction is less important. Biorisk I'm less sure about - non-AI mediated extinction from bio seems very unlikely, but what would a GCR pandemic do to future values? Probably ~neutral in expectation, but plausibly it could lead to the demise of liberal democratic institutions (bad), or to a post-recovery world that is more scared and committed to global cooperation to prevent that recurring (good).
Lizka
15
0
0
1

I wrote a Twitter thread that summarizes this piece and has a lot of extra images (I probably went overboard, tbh.) 

I kinda wish I'd included the following image in the piece itself, so I figured I'd share it here:

As AI capabilities rise, AI systems will be responsible for a growing fraction of relevant work

I just want to say that I really appreciated this post — it came at exactly the right time for me and I've referenced it several times since you shared it. 

4
Joe_Carlsmith
Very glad to hear it, Lizka :) -- and thanks for letting me know. 

Follow-up: 

Quick list of some ways benchmarks might be (accidentally) misleading[1]

  1. Poor "construct validity"[2]( & systems that are optimized for the metric)
    1. The connection between what the benchmark is measuring and what it's trying to measure (or what people think it's measuring) is broken. In particular:
    2. Missing critical steps
      1. When benchmarks are trying to evaluate progress on some broad capability (like "engineering" or "math ability" or "planning"), they're often testing specific performance on meaningfully simpler tasks. Performance on those t
... (read more)

TLDR: Notes on confusions about what we should do about digital minds, even if our assessments of their moral relevance are correct[1]

I often feel quite lost when I try to think about how we can “get digital minds right.” It feels like there’s a variety of major pitfalls involved, whether or not we’re right about the moral relevance of some digital minds.

Digital-minds-related pitfalls in different situations

Reality ➡️

Our perception ⬇️

These digital minds are (non-trivially) morally relevant[2]These digital minds are not morally relevant
We see thes
... (read more)
3
Alfredo Parra 🔸
@Lucius Caviola has also written about this topic, e.g. Will disagreement about AI rights lead to societal conflict? And my two cents on why I don't think we should worry about digital sentience (plus the slicing problem). :)
2
Toby Tremlett🔹
Thanks for writing this! I often think about this question with almost that exact numbered grid. And now I can cite your quick take :) I'm especially interested in outcome (2) (see this old post).  PS- this outcome is also known as "Disneyland with no children".

Thanks for saying this! 

I’m trying to get back into the habit of posting more content, and aiming for a quick take makes it easier for me to get over perfectionism or other hurdles (or avoid spending more time on this kind of thing than I endorse). But I’ll take this as a nudge to consider sharing things as posts more often. :)

3
NickLaing
Love it! I like that idea, and if its a lower bar psychologically to post in quicktakes that makes sense :).
Lizka
82
10
0
16
1

When I try to think about how much better the world could be, it helps me to sometimes pay attention to the less obvious ways that my life is (much) better than it would have been, had I been born in basically any other time (even if I was born among the elite!).

So I wanted to make a quick list of some “inconspicuous miracles” of my world. This isn’t meant to be remotely exhaustive, and is just what I thought of as I was writing this up. The order is arbitrary.

1. Washing machines

It’s amazing that I can just put dirty clothing (or dishes, etc.) into a ... (read more)

3
akash 🔸
I love this write up. Re point 2 — I sincerely think we are in the golden age of media, at least in ~developed nations. There has never been a time where any random person could make music, write up their ideas, or shoot an independent film and make a living out of it! The barrier to entry is so much lower, and there are typically no unreasonable restrictions on the type of media we can create (I am sure medieval churches wouldn't be fans of heavy metal). If we don't mess up our shared future, all this will only get better. Also, I feel this should have been a full post and not a quick note.
9
Mo Putera
I like this; feels like a more EA-flavored version of Gwern's My Ordinary Life: Improvements Since the 1990s. 
4[anonymous]
Great image selection!

Notes on some of my AI-related confusions[1]

It’s hard for me to get a sense for stuff like “how quickly are we moving towards the kind of AI that I’m really worried about?” I think this stems partly from (1) a conflation of different types of “crazy powerful AI”, and (2) the way that benchmarks and other measures of “AI progress” de-couple from actual progress towards the relevant things. Trying to represent these things graphically helps me orient/think. 

First, it seems useful to distinguish the breadth or generality of state-of-the-art AI ... (read more)

4
Denkenberger🔸
I found these visualizations very helpful! I think of AGI as the top of your HLAI section: human level in all tasks. Life 3.0 claimed that just being superhuman at AI coding would become super risky (recursive self improvement (RSI)). But it seems to me it would need to be ~human level at some other tasks as well like planning and deception to be super risky. Still, that could be relatively narrow overall.
4
Lizka
Follow-up:  Quick list of some ways benchmarks might be (accidentally) misleading[1] 1. Poor "construct validity"[2]( & systems that are optimized for the metric) 1. The connection between what the benchmark is measuring and what it's trying to measure (or what people think it's measuring) is broken. In particular: 2. Missing critical steps 1. When benchmarks are trying to evaluate progress on some broad capability (like "engineering" or "math ability" or "planning"), they're often testing specific performance on meaningfully simpler tasks. Performance on those tasks might be missing key aspects of true/real-world/relevant progress on that capability. 3. Besides the inherent difficulty of measuring the right thing, it's important to keep in mind that systems might have been trained specifically to perform well on a given benchmark. 1. This is probably a bigger problem for benchmarks that have gotten more coverage. 4. And some benchmarks have been designed specifically to be challenging for existing leading models, which can make new/other AI systems appear to have made more progress on these capabilities (relative to the older models) than they actually are. 1. We're seeing this with the "Humanity's Last Exam" benchmark. 5. ...and sometimes some of the apparent limitations are random or kinda fake, such that a minor improvement appears to lead to radical progress 2. Discontinuous metrics: (partial) progress on a given benchmark might be misleading. 1. The difficulty of tasks/tests in a benchmark often varies significantly (often for good reason), and reporting of results might explain the benchmark by focusing on its most difficult tests instead of the ones that the model actually completed. 1. I think this was an issue for Frontier Math, although I'm not sure how much strongly to discount some of the results as a result. 2. -> This (along with issues like 1b above, which can lead to saturation) is part of

Lizka might not work for the forum anymore, but I would have thought she could see this is way too deep and good for a quicktake!

I don't think the situation is actually properly resolved by the recent waivers (as @Garrison points out below). See e.g. this thread and the linked ProPublica article. From the thread:

There is mass confusion and fear, both in and outside government. The aid organizations say they either don’t know how to get a waiver exempting them from Trump’s order or have no idea if/when theirs might get approved.

And from the article:

Despite an announcement earlier this week ostensibly allowing lifesaving operations to continue, those earlier orders have not been resci

... (read more)
2
Rebecca
What further actions would Trump/Rubio need to take to properly execute the waiver?

As a datapoint: despite (already) agreeing to a large extent with this post,[1] IIRC I answered the question assuming that I do trust the premise. 


Despite my agreement, I do think there are certain kinds of situations in which we can reasonably use small probabilities. (Related post: Most* small probabilities aren't pascalian, and maybe also related.) 


More generally: I remember appreciating some discussion on the kinds of thought experiments that are useful, when, etc. I can't find it quickly, but possible starting points could be this LW po... (read more)

That makes sense and is roughly how I was interpreting what you wrote (sorry for potentially implying otherwise in my comment) — this is still a lot more positive on peacekeeping than I was expecting it to be :) 

3
Lauren Gilbert
Yes, I thought that was what you meant but wanted to be clear - I very much don't think that GiveWell should start recommending the UN.  ;)

Before looking at what you wrote, I was most skeptical of the existence of (plausibly) cost-effective interventions on this front. In particular, I had a vague background view that some interventions work but are extremely costly (financially, politically, etc.), and that other interventions either haven't been tried or don't seem promising. I was probably expecting your post to be an argument that we/most people undervalue the importance of peace (and therefore costly interventions actually look better than they might) or an argument that there are some n... (read more)

5
Lauren Gilbert
I should note - I don't think peacekeeping is anywhere near as cost-effective as GiveWell's top interventions! My (very quick, rough) BOTEC on peacekeeping in 2022 had it about half as good as GiveDirectly (see the civil conflict shallow and associated BOTEC).  Peacekeeping should not be an EA cause area.  Getting the UN to focus more on peacekeeping and less on other functions?  That might pencil, since it's leveraged (though I am very uncertain on that).

Addendum to the post: an exercise in giving myself advice

The ideas below aren't new or very surprising, but I found it useful to sit down and write out some lessons for myself. Consider doing something similar; if you're up for sharing what you write as a comment on this post, I'd be interested and grateful.


 (1) Figure out my reasons for working on (major) projects and outline situations in which I would want myself to leave, ideally before getting started on the projects

I plan on trying to do this for any project that gives me any (ethical) doubts, a... (read more)

2
OscarD🔸
Choosing 'and' or 'or' feels important here since they seem quite different! Maybe our rough model should be cause-for-introspection = ethical qualms * length of project
Lizka
63
15
0

A note on how I think about criticism

(This was initially meant as part of this post,[1] but while editing I thought it didn't make a lot of sense there, so I pulled it out.)

I came to CEA with a very pro-criticism attitude. My experience there reinforced those views in some ways,[2] but it also left me more attuned to the costs of criticism (or of some pro-criticism attitudes). (For instance, I used to see engaging with all criticism as virtuous, and have changed my mind on that.) My overall takes now aren’t very crisp or easily summarizable,... (read more)

8
Ben_West🔸
I would be excited about this and have wondered for a while if we should have EA awards. This Washington post article brought the idea to my mind again:
Lizka
44
7
0
4
3

A note on mistakes and how we relate to them

(This was initially meant as part of this post[1], but I thought it didn't make a lot of sense there, so I pulled it out.)

“Slow-rolling mistakes” are usually much more important to identify than “point-in-time blunders,”[2] but the latter tend to be more obvious.

When we think about “mistakes”, we usually imagine replying-all when we meant to reply only to the sender, using the wrong input in an analysis, including broken hyperlinks in a piece of media, missing a deadline, etc. I tend to feel pretty horrible ... (read more)

6
MichaelDickens
After reading your post, I wasn't sure you were right about this. But after thinking about it for a few minutes, I can't come up with any serious mistakes I've made that were "point-in-time blunders". The closest thing I can think of is when I accidentally donated $20,000 to the GiveWell Community Foundation instead of The Clear Fund (aka GiveWell), but fortunately they returned the money so it all worked out.
Lizka
37
7
1
2

I'm going to butt in with some quick comments, mostly because:

  • I think it's pretty important to make sure the report isn't causing serious misunderstandings 
  • and because I think it can be quite stressful for people to respond to (potentially incorrect) criticisms of their projects — or to content that seem to misrepresent their project(s) — and I think it can help if someone else helps disentangle/clarify things a bit. (To be clear, I haven't run this past Linch and don't know if he's actually finding this stressful or the like. And I don't want to disc
... (read more)

I'd suggest using a different term or explicitly outlining how you use "expert" (ideally both in the post and in the report, where you first use the term) since I'm guessing that many readers will expect that if someone is called "expert" in this context, they're probably "experts in EA meta funding" specifically — e.g. someone who's been involved in the meta EA funding space for a long time, or someone with deep knowledge of grantmaking approaches at multiple organizations. (As an intuition pump and personal datapoint, I wouldn't expect "experts" in the c... (read more)

I know Grace has seen this already, but in case others reading this thread are interested: I've shared some thoughts on not taking the pledge (yet) here.[1]

Adding to the post: part of the value of pledges like this comes from their role as a commitment mechanism to prevent yourself from drifting away from values and behaviors that you endorse. I'm not currently worried about drifting in this way, partly because I work for CEA and have lots of social connections to extremely altruistic people. If I started working somewhere that isn't explicitly EA-oriented... (read more)

2
GraceAdams🔸
Oh I thought I responded to this already! I'd like to say that people often have very good reasons for not pledging, that are sometimes visible to us, and other times not - and no one should feel bad for making the right choice for themselves!  I do of course think many more people in our community could take the GWWC Pledge, but I wouldn't want people to do that at the expense of them feeling comfortable with making that commitment. We should respect other people's journeys, lifestyles and values in our pursuits to do good. And thanks Lizka for sharing your previous post in this thread too! Appreciate you sharing your perspective!
6
GraceAdams🔸
I think it's pretty unacceptable to be rude or unkind to anyone who hasn't take a pledge with GWWC. Everyone is on their own journey and should do what is right for them. I would be disappointed to hear of pledgers who are acting in a manner that's unkind to non-pledgers. I second Liza's request here to ask people who are being uncharitable or unkind about the decisions of others around taking a pledge to refrain from doing so.  I think it's acceptable to politely ask people if they'd welcome a discussion about reasons they should consider taking a pledge, but if there's no interest, to let the conversation go. One of the reasons I love the GWWC Community, is that people tend to be very kind and welcoming, and I would hate to see that change. (also thanks Liza for sharing your previous post here, too!)

As a quick update: I did not in fact share two posts during the week. I'll try to post another "DAW post" (i.e. something from my drafts, without spending too much time polishing it) sometime soon, but I don't endorse prioritizing this right now and didn't meet my commitment. 

Answer by Lizka10
2
0

Not sure if this already exists somewhere (would love recommendations!), but I'd be really excited to see a clear and carefully linked/referenced overview or summary of what various agriculture/farming ~lobby groups do to influence laws and public opinion, and how they do it (with a focus on anything related to animal welfare concerns). This seems relevant.

Just chiming in with a quick note: I collected some tips on what could make criticism more productive in this post: "Productive criticism: what could help?"

I'll also add a suggestion from Aaron: If you like a post, tell the author! (And if you're not sure about commenting with something you think isn't substantive, you can message the author a quick note of appreciation or even just heart-react on the post.) I know that I get a lot out of appreciative comments/messages related to my posts (and I want to do more of this myself). 

8
Jason
I think it would be particularly valuable to tell the author (and the rest of the Forum) how you think reading the post created impact. More concrete / personal examples might include: I am more likely to donate to X / using Y strategy because I read this post, or I will be more likely to advise people who come to me for career advice to do Z. I think many potential authors may be wondering whether the time spent writing actually produces impact in comparison to the counterfactual uses of their time, so feedback on that point should be helpful.

I'll commit to posting a couple of drafts. Y'all can look at me with disapproval (or downvote this comment) if I fail to share two posts during Draft Amnesty Week. 

2
Lizka
As a quick update: I did not in fact share two posts during the week. I'll try to post another "DAW post" (i.e. something from my drafts, without spending too much time polishing it) sometime soon, but I don't endorse prioritizing this right now and didn't meet my commitment. 
6
NickLaing
Amazing. Ready to strong downvote this comment at even the hint of only one draft by the 17th March. We'll see if a ulysses-stylye pact as weak as a few negative Karma on the EA will be enough motivation ;).
Answer by Lizka21
3
0

I'm basically always interested in potential lessons for EA/EA-related projects from various social movements/fields/projects.

Note that you can find existing research that hasn't been discussed (much) on the Forum and link-post it (I bet there's a lot of useful stuff out there), maybe with some notes on your takeaways. 

Example movements/fields/topics: 

  • Environmentalism — I've heard people bring up the environmentalist/climate movement a bunch in informal discussions as an example for various hypotheses, including "movements splinter/develop highly
... (read more)
Answer by Lizka5
1
0

I'd love to see two types of posts that were already requested in the last version of this thread:

  • From Aaron: "More journalistic articles about EA projects. [...] Telling an interesting story about the work of a person/organization, while mixing in the origin story, interesting details about the people involved, photos, etc."
  • From Ben: "More accessible summaries of technical work." (I might share some ideas for technical work I'd love to see summarized later.)

I really like this post and am curating it (I might be biased in my assessment, but I endorse it and Toby can't curate his own post). 

A personal note: the opportunity framing has never quite resonated with me (neither has the "joy in righteousness" framing), but I don't think I can articulate what does motivate me. Some of my motivations end up routing through something ~social. For instance, one (quite imperfect, I think!) approach I take[1] is to imagine some people (sometimes fictional or historical) I respect and feel a strong urge to be the ... (read more)

Thanks for sharing this! I'm going to use this thread as a chance to flag some other recent updates (no particular order or selection criteria — just what I've recently thought was notable or recently mentioned to people): 

  1. California proposes sweeping safety measure for AI — State Sen. Scott Wiener wants to require companies to run safety tests before deploying AI models. (link goes to "Politico Pro"; I only see the top half)
    1. Here's also Senator Scott Wiener's Twitter thread on the topic (note the endorsements)
    2. See also the California effect
  2. Trump: AI ‘m
... (read more)

I don't actually think you need to retract your comment — most of the teams they used did have (at least some) biological expertise, and it's really unclear how much info the addition of the crimson cells adds. (You could add a note saying that they did try to evaluate this with the additional of two crimson cells? In any case, up to you.)

(I will also say that I don't actually know anything about what we should expect about the expertise that we might see on terrorist cells planning biological attacks — i.e. I don't know which of these is actually appropriate.)

2
Agustín Covarrubias 🔸
Changed it to a note. As for the latter, my intuition is that we should probably hedge for the full spectrum, from no experience to some wet bio background (but the case where we get an expert seems much more unlikely).
Load more