All of Jonas Hallgren 🔸's Comments + Replies

I enjoyed reading this and yet I find that in the practice of higher ambition there are some specific pitfalls that I still haven't figured out my way around. 

If you've every worked a 60-70 hour work week and done it for a longer period of time, you can notice a narrowing characteristic of experience, it is as if you have blinders to what is not within your stated goals or the project you're working on. (I like to call this compression) With some of my more ambitious friends who do this more often, I find that they sometimes get lost in what they're w... (read more)

8
cb
I wonder if having scheduled downtime to rest, reflect, and decide your next moves would work here? Intuitively, it seems like "sprint on a goal for a quarter, take a week (or however long) to reflect and red-team your plans for the next quarter, then sprint on the new plans, etc" would minimise a lot of the downside, especially if you're already working on pretty well-scoped, on-point projects. (I think committing to a "tour of duty" on a job/project, and then some time to reflect and evaluate your next steps, has similar benefits.) (I can see how you might want more/longer reflective periods if you're choosing between more speculative, sign-uncertain projects.)

Are you building these things on ATProtocol (Bluesky) or where are you building it right now? I feel like there's quite a nice movement happening there with some specific tools for this sort of thing. (I'm curious because I'm also trying to build some stuff more on the deeper programming level but I'm currently focusing on open-source bridging and recommendation algorithms like pol.is but for science and it would be interesting to know where other people are building things.)

If you don't know about the ATProtocol gang, some things I enjoy here are:
- https:... (read more)

2
mako yass
I don't think atproto is really a well designed protocol * No private records yet, so can't really build anything you'd wanna live in on it. * Would an agenty person respond to this situation by taking atproto and inventing their own private record extension for it and then waiting for atproto to catch up with them? Maybe. But also: * The use of dns instead of content-addressing for record names is really ugly, since they're already using content-addressing elsewhere, so using dns is just making it hard to make schema resolution resilient, and it prohibits people who aren't sysadmins from publishing schemas. (currently only sysadmin types could need a schema published, but it's lame that that's the case. Anyone should be able to publish data. Anyone can define a type.) In theory you can still do this stuff (provide a special domain that's an ipfs gateway or something and treat records that use that as having content-addressed specs), but at that point it seems like you're admitting that atproto was mostly just a mistake? * The schema system isn't a good type system (lacks generics). They probably don't think of it as a type system (but it is, and should be). And the ecosystem currently has nothing to offer, not really * Would anyone benefit from integrating closely with bsky? * There are people who've set up their blogs so that bsky replies to links to their post show up in the comments, for instance, but I'd guess most actual bloggers should (and perhaps already do) actively not want to display the replies from character limited forums, because, you know, they tend to be dismissive, not cite sources, not really be looking for genuine conversation, etc. * You could use a person's bsky following as a general purpose seed list for comment moderation I guess. But with a transitive allow list this isn't really needed. * I don't expect meaningful amounts of algorithmic choice to just happen. I'd guess that training good sequential recommender systems is

Firstly, that is if you think that it isn't inevitable and that it is possible to stop or slow down, if nuclear was going to be developed anyway, that changes the calculus. Even if that is the case there's also this weird thing within human psychology where if you can point out a positive vision of something, it is often easier for people to kind of get it? 

"Don't do this thing" is often a lot worse than saying something like, could you do this specific thing instead when it comes to convincing people of things. This is also true for specific therapeu... (read more)

I would very much be curious about mechanisms for the first point you mentioned! 

For 11, I would give a little bit of pushback related to your building as a sport team metaphor as I find them a bit discongruant with each other? 

Or rather the degree of growth mindset that is implied in 11th seems quite bad based on best practices within things like sport psychology and general psychology? The existent frame is like you're either elite or you're not gonna make it. I would want the frame to be like "it's really hard to become a great football player... (read more)

Firstly, great post thanks for writing it!

Secondly, with regards to the quantification section:

Putting numbers on the qualities people have feels pretty gross, which is probably why using quantification in hiring is rather polarising. On the one hand, there’s some line of thinking that the different ways in which people are well and ill suited to particular roles isn’t quantifiable and if you try to quantify it you’ll just be introducing bias. On the other hand, people in favour of quantification tend to strongly recommend that you stick exactly to the ran

... (read more)

Very very well put. 

I became quite emotional when reading this because I resonated with it quite strongly. I've been in some longer retreats practicing the teachings in Seeing That Frees and I've noticed the connections between EA and Rob Burbea's way of seeing things but I haven't been able to express it well. 

I think that there's a very beauitful deepening of a seeing of non-self when acting impartialy. One of the things that I really like about applying this to EA is that you often don't see the outcomes of your actions. This is often seen as ... (read more)

3
kuhanj
Thank you for the kind words Jonas! Your comment reminded me of another passage from one of my favorite Rob talks, Selflessness and a Life of Love:

The question that is on every single EAs mind is, of course, what about huel or meal replacements? I've been doing huel+supplements for a while now instead of meat and I want to know if you believe this to be suboptimal and if so to what extent? Nutrition is annoyingly complex and so all I know for sure is like protein=good, cal in=cal out and minimize sugar (as well as some other things) and huel seems to tick all the boxes? I'm probably missing something but I don't know what so if you have an answer, please enlighten me!

This one hit close to home (pun not intended). 

I've been thinking about this choice for a while now. There's the obvious network and work benefits in living in an EA Hub yet in my experience there's also the benefit of a slower pace leading to more time to think and reflect and develop my own writing and opinions on things which is easier to get when not in a hub.

Yet in AI safety (where I work) all of the stuff is happening in the Bay and London and mostly the Bay. For the last 3 years people have constantly been telling me "Come to the Bay, bro. It w... (read more)

6
ElliotTep
Wow that's gotta be one of the fastest forum post to plan changes on record. I'm glad to hear this resolved what sounds like a big and tough question in your life. As I mentioned in the post, I do think stints in hubs can be a great experience.

Uncertain risk. AI infrastructure seems really expensive. I need to actually do the math here (and I haven’t! hence this is uncertain) but do we really expect growth on trend given the cost of this buildout in both chips and energy? Can someone really careful please look at this?

 

https://www.lesswrong.com/users/vladimir_nesov <- Got a bunch of stuff on energy calculations and similar required for AI companies, especially the 2028 post, some very good analysis of these things imo.

I think it is a bit like the studies on what makes people able to handle adversity well, it's partly about preparation and ensuring that the priors people bring into the systems are equipped to handle the new attack vectors that this transition provides to our collective epistemics. 

So I think we need to create some shared sources of trust that everyone can agree on and establish those before the TAI transition if we want things to go well.

1
Ben Norman
Thanks for your comment! I agree studying how people handle adversity is an important direction, but I think that creating “shared sources of trust everyone can agree on” would be hard to do in practice. What would you concretely imagine this looking like?

I'm curious about the link that goes to AI-enabled coups and it isn't working, could you perhaps relink it?

2
Jackson Wagner
Sorry about that!  I think I just intended to link to the same place I did for my earlier use of the phrase "AI-enabled coups", namely this Forethought report by Tom Davidson and pals, subtitled "How a Small Group Could Use AI to Seize Power": https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power But also relevant to the subject is this Astral Codex Ten post about who should control an LLM's "spec": https://www.astralcodexten.com/p/deliberative-alignment-and-the-spec The "AI 2027" scenario is pretty aggressive on timelines, but also features a lot of detailed reasoning about potential power-struggles over control of transformative AI which feels relevant to thinking about coup scenarios.  (Or classic AI takeover scenarios, for that matter. Or broader, coup-adjacent / non-coup-authoritarianism scenarios of the sort Thiel seems to be worried about, where instead of getting taken over unexpectedly by China, Trump, or etc, today's dominant western liberal institutions themselves slowly become more rigid and controlling.) For some of the shenanigans that real-world AI companies are pulling today, see the 80,000 Hours podcast on OpenAI's clever ploys to do away with its non-profit structure, or Zvi Mowshowitz on xAI's embarrassingly blunt, totally not-thought-through attempts to manipulate Grok's behavior on various political issues (or a similar, earlier incident at Google).

Besides the point that "shoddy toy models" might be emotionally charged, I just want to point out that accelerating progress majorly increases variance and unknown unknowns? The higher energy a system is and the more variables you have the more chaotic it becomes. So maybe an answer is that a agile short-range model is the best? Outside view it in moderation and plan with the next few years being quite difficult to predict?

You don't really need another model to disprove an existing one, you might as well point out that we don't know and that is okay too.

Yeah, I think you're right and I also believe that it can be a both and? 

You can have a general non-profit board and at the same time have a form of representative democracy going on which seems the best we can currently do for this?

I think it is fundamentally about a more timeless trade-off between hierarchical organisations that generally are able to act with more "commander's intent" versus democratic models that are more of a flat voting model. The democratic models suffer when there is a lot of single person linear thinking involved but do well a... (read more)

4
Patrick Gruban 🔸
I can see democratic models providing value, but the practical implementation is tricky. I can only speak from my experience in EA Germany, where member engagement in national-level strategy and participation in the national community seemed much lower than what I experience on the international level (in this forum, for example) or even at the city level at times. I would be more excited about either local structures (cities or small regions with fewer than 10 million people) or larger structures (sub-continents, professional groups, etc.) where people truly form a community in the sense that they see each other in person, or where there is a large enough body to allow for meaningful participation in democratic processes. 

Yeah for sure, I think the devil might be in the details here around how things are run and what the purpose of the national organisation is. Since Sweden and Norway have 8x less of a population than germany I think the effect of a "nation-wide group" might be different?

In my experience, I've found that EA Sweden focuses on and provides a lot of the things that you listed so I would be very curious to hear what the difference between a local and national organisation would be? Is there a difference in the dynamics of them being motivated to sustain themselves because of the scale? 

You probably have a lot more experience than me in this so it would be very interesting to hear!

I like that decomposition. 

There's something about a prior on having democratic decision making as part of this because it allows for better community engagement usually? Representation often leads to feelings of inclusion and whilst I've only dabbled in the sociology here it seems like the option of saying no is quite important for members to feel heard? 

My guess would be that the main pros of having democratic deliberation doesn't come from when the going is normal but rather as a resillience mechanism? Democracies tend to react late to major c... (read more)

4
Patrick Gruban 🔸
Perhaps, but I can also imagine that a hand-selected nonprofit board may be able to spot risks and react to them better than a board voted in an assembly. The coordination function of an assembly in trying to fill specific board roles seems lower than if a smaller group of existing board members can discuss it.

I think I went through it in Spring 2021? I remember discussing it then as one of the advanced optional topics, maybe around steering versus rowing and that the discussion went into that? I can't remember it more clearly than that though.

First and foremost, I think the thoughts expressed here make sense and this comment is more just expressing a different perspective, not necessarily disagreeing.

I wanted to bring up an existing framework for thinking about this from Raghuram Rajan's "The Third Pillar," which provides economic arguments for why local communities matter even when they're less "efficient" than centralized alternatives.

The core economic benefits of local community structures include:

  • Information advantages: Local groups understand context that centralized organizations miss
  • Adap
... (read more)
2
Patrick Gruban 🔸
I agree with the benefits of local community structures. However, I don't believe that national EA groups can offer as much as informal local groups. I help manage both formal and informal networks of EA (adjacent) individuals in Munich, and there, I see these points much more clearly. Running a coworking space, hosting in-person events, convening private meetings, and having one-on-ones seem like activities that would fit your list.
2
Jason
If I'm reading Patrick's comment correctly, there are two different ideas going on: * The democratic approach requires greater overhead (e.g., "the overhead for organizing a general assembly") without producing better results to justify the extra overhead * Fewer / geographically broader orgs would have greater efficiency for ~the usual reasons we might think larger orgs might do better than smaller ones These effects should be, in theory, somewhat separate -- one could envision a nationally focused org without membership/democracy, or a big transnational group with it. Do you think your list of advantages is more about localness or more about being democratic?
5
Jacob Watts🔸
  I am sympathetic to this argument vibes wise and I thought this was an elegant numerate utilitarian case for it. Part of my motivation is that I think it would be good if a lot of EA-ish values were a lot more mainstream. Like, I would even say that you probably get non-linear returns to scale in some important ways. You kind of need a critical mass of people to do certain things.  It feels like, necessarily, these organizations would also be about providing value to the members as well. That is a good thing. I think there is something like a "but what if we get watered down too much" concern latent here. I can kind of see how this would happen, but I am also not that worried about it. The tent is already pretty big in some ways. Stuff like numerate utilitarianism, empiricism, broad moral circles, thoughtfulness, tough trade-offs doesn't seem in danger of going away soon. Probably EA growing would spread these ideas rather than shrink them. Also, I just think that societies/people all over the world could significantly benefit from stronger third pillars and that the ideal versions of these sorts of community spaces would tend to share a lot of things in common with EA.  Picture it. The year is 2035 (9 years after the RSI near-miss event triggered the first Great Revolt). You ride your bitchin' electric scooter to the EA-adjacent community center where you and your friends co-work on a local voter awareness campaign, startup idea, or just a fun painting or whatever. An intentional community.  That sounds like a step towards the glorious transhumanist future to me, but maybe the margins on that are bad in practice and the community centers of my day dreams will remain merely EA-adjacent. Perhaps, I just need to move to a town with cooler libraries. I am really not sure what the Dao here is or where the official EA brand really fits into any of this. 

This is very nice! 

I've been thinking that there's a nice generalisable analogy between bayesian updating and forecasting. (It is quite no shit when you think about it but it feels like people aren't exploiting it?)

I'm doing a project on simulating a version of this idea but in a way that utilizes democratic decision making called Predictive Liquid Democracy (PLD) and I would love to hear if you have any thoughts on the general setup. It is model parameterization but within a specific democratic framing.

PLD is basically saying the following:

What if we... (read more)

Some people might find that this post is written from a place of agitation which is fully okay. I think that even if you do there are two things that I would want to point out as really good points:

  1. A dependence on funders and people with money as something that shapes social capital and incentives, therefore thought in itself. We should therefore be quite vary of the effect that has on people, this can definetely be felt in the community and I think it is a great point.
  2. That the karma algorithm could be revisited and that we should think about what incentiv
... (read more)
1
Yi-Yang
Do you remember which in-depth fellowship's syllabus talks about elitism in EA? I've seen a few versions the syllabus before but I don't think I have encountered a piece of reading on elitism. No worries if you don't!
1
Maxim Vandaele
Thank you for the constructive reply. As my friend Bob has proposed, one of the most promising proposals is to run an EA organization as a workers' cooperative (see Bob's post about the benefits of cooperatives). It could bring about greater trust in EA organizations and possibly better decision-making by those organizations.

I felt that this post might be relevant for longtermism and person affecting views so I had claude write up a quick report on that:

In short: Rejecting the SWWM 💸11% pledge's EV calculation logically commits you to person-affecting views, effectively transforming you from a longtermist into a neartermist.

Example: Bob rejects investing in a $500 ergonomic chair despite the calculation showing 10^50 * 1.2*10^-49 = 12 lives saved due to "uncertainty in the probabilities." Yet Bob still identifies as a longtermist who believes we should value future generation... (read more)

First and foremost, I'm low confidence here. 

I will focus on x-risk from AI and I will challenge the premise of this being the right way to ask the question.

What is the difference between x-risk and s-risk/increasing the value of futures? When we mention x-risk with regards to AI we think of humans going extinct but I believe that to be a shortform for wise compassionate decision making. (at least in the EA sphere) 

Personally, I think that x-risk and good decision making in terms of moral value might be coupled to each other. We can think of our ... (read more)

1
Chris Clay🔸
I've heard this argument before, but I find it un-compelling in its tractability. If we don't go extinct, its likely to be a silent victory; most humans on the planet won't even realise it happened. Individual humans working on X-risk reduction will probably only impact the morals of people around them.

First and foremost, I agree with the point. I think looking at this especially from a lens of transformative AI might be interesting. (Coincidentally this is something I'm currently doing using ABMs with LLMs)

You probably know this one but here's a link to a cool project: https://effectiveinstitutionsproject.org/

Dropping some links below, I've been working on this with a couple of people in Sweden for the last 2 years, we're building an open source platform for better democratic decision making using prediction markets:

https://digitaldemocracy.world/flowba... (read more)

I guess a random thought I have here is that you would probably want video and you would probably want it to be pretty spammable so you have many shots at it. Looking at twitter we already see like a large amounts of bots around commenting on things which is like a text deepfake.

Like I can see in a year or so when SORA is good enough that creating a short form stabel video is easy we will see a lot more manipulation of voters through various social media through deepfakes. 

(I don't think the tech is easy enough to use yet for it to be painless to do i... (read more)

FWIW, I find that if you analyze places where we've successfully aligned things in the past (social systems or biology etc.) you find that the 1th and 2nd types of alignment really don't break down in that way. 

After doing Agent Foundations for a while I'm just really against the alignment frame and I'm personally hoping that more research in direction will happen so that we get more evidence that other types of solutions are needed. (e.g alignment of complex systems such as has happened in biology and social systems in the past)

2
SiebeRozendal
That sounds like [Cooperative AI](https://www.cooperativeai.com/post/new-report-multi-agent-risks-from-advanced-ai)  https://www.cooperativeai.com/post/new-report-multi-agent-risks-from-advanced-ai 

FWIW, I completely agree with what you're saying here and I think that if you seriously go into consciousness research and especially for what we westerners more label as a sense of self rather than anything else it quickly becomes infeasible to hold a position that the way we're taking AI development, e.g towards AI agents will not lead to AIs having self-models. 

For all matters and purposes this encompasses most theories of physicalist or non-dual theories of consciousness which are the only feasible ones unless you want to bite some really sour app... (read more)

I'm not a career councellor so take everything with a grain of salt but you did publically post this asking for unsolicited advice, so here you go! 

So, more directly if you're thinking of EA as a community that needs specific skills and you're wondering what to do, your people management skills, strategy & general leadership skills are likely to be high in demand from other organisations: https://forum.effectivealtruism.org/posts/LoGBdHoovs4GxeBbF/meta-coordination-forum-2024-talent-need-survey

Someone else mentioned that enjoyment can be highly or... (read more)

So I'll just give some reporting on a vibe I've been feeling on the forum.

I feel a lot more comfortable posting on LessWrong compared to the EA forum because it feels like there's a lot more moral outrage here? Like if I go back 3 or 4 years I felt that the forum was a lot more open to discussing and exploring new ideas. There's been some controversies recently around meat-eater problem stuff and similar and I can't help but just feel uncomfortable posting stuff with how people have started to react? 

I like the different debate weeks as I think it set... (read more)

I want to preface that I don't have a strong opinion here, just some curiosity and a question.

If we are focusing on second order effects wouldn't it make sense to bring up something like moral circle expansion and its relation to ethical and sustainable living over time as well?

From a long-term perspective, I see one of the major effects of global health being better decision making through moral circle expansion.

My question to you is then what time period you're optimising for? Does this matter for the argument?

6
Vasco Grilo🔸
Thanks for the comment, Jonas. I think focussing on impartial welfare by accounting for effects on farmed animals is better for moral circle expansion than completely ignoring effects on animals, which is currently the stardard in assessing interventions helping humans. In principle, I care about effects across all space and time. In practice, I think the effects after 100 years are negligible. I guess the ratio between harms to farmed animals and benefits to humans globally, and in China, India and Nigeria will initially increase as the consumption per capita of animal-based food increases, but then decrease as the consumption per capita of animal-based food stagnates or decreases, and the conditions of animals improve.

Thank you for that substantive response, I really appreciate it! It was also very nice that you mentioned the Turner et.al definitions, I wasn't expecting that. 

(Maybe write a post on that? There's a comment that mentions uptake from major players in the EA ecosystem and maybe if you acknowledge you understand the arguments they would be more sympathetic? Just a quick thought but it might be worth engaging there a bit more?)

I just wanted to clarify some of the points I was trying to make yesterday as I do realise that they didn't all get across as I w... (read more)

Thank you for this post David! 

I've from time to time engaged with my friends in discussion about your criticisms of longtermism and some existential risk calculations. I found that this summary post of your work and interaction calrifies my perspective on the general "inclination" that you have in engaging with the ideas, one that seems like a productive one! 

Sometimes, I felt that it didn't engage with some of the core underlying claims of longtermism and exisential risk which did annoy me. 

I want to respect the underlying time spend assym... (read more)

9
David Thorstad
Thanks Jonas! I really appreciate your constructive engagement.  I’m not very sympathetic to pure time preference – in fact, most philosophers hate pure time preference and are leading the academic charge against it. I’m also not very sympathetic to person-affecting views, though many philosophers whom I respect are more sympathetic. I don’t rely on either of these views in my arguments, because I think they are false. In general, my impression on the topic of advancing progress is that longtermists have not yet identified plausible interventions which could do this well enough to be worth funding. There has certainly been a lot written about the value of advancing progress, but not many projects have been carried out to actually do this. That tends to suggest to me that I should hold off on commenting about the value of advancing progress until we have more concrete proposals on the table. It also tends to suggest that advancing progress may be hard.  I think that covers most of the content of Toby’s posts (advancing progress and temporal discounting). Perhaps one more thing that would be important to stress is that I would really like to see Oxford philosophers following the GPI model of putting out smaller numbers of rigorous academic papers rather than larger numbers of forum posts and online reports. It’s not very common for philosophers to engage with one another on internet forums – such writing is usually regarded as a form of public outreach. My first reaction to EA Forum posts by Toby and others would be that I’d like to see them written up as research papers. When ideas are written up online before they are published in academic venues, we tend to be skeptical that the research has actually been done or that it would hold up to scrutiny if it were. On power-seeking, one of the most important features of academic papers is that they aim to tackle a single topic in depth, rather than a broad range of topics in less depth. So for example, my power-seeki

I've found a lot of my EA friends falling into this decision paralysis so thank you for this post, I will link this to them! 

3
ElliotTep
Glad to hear. The goal was very much to write the kind of post people want to reference when chatting to friends who are making this mistake.

I just did different combinations of the sleep supplements, you still get the confounder effects but it removes some of the cross-correlation. So Glycine 3 days, no magnesium followed by magnesium 3 days, no glycine e.t.c. It's not necessarily going to give you a high accuracy but you can see if it works or not and a rough effect size

I use bearable for 3 months at a time to get a picture of what is currently working. You can track effect sizes of supplements in sleep quality for example if you also have a way of tracking your sleep. 

Funnily enough, I noticed there were a bunch of 80/20 stuff in my day through using bearable. I found doing a cold shower, loving kindness meditation in the morning and getting sunlight in the morning were like a difference of 30% in energy and enjoyment so I now do these religiously and it has worked wonders. (I really like bearable for these sorts of experiments.)

2
Jessica McCurdy🔸
This is super interesting! How do you do the experiments? Do you change one thing at a time and track?

Sorry for not noticing the comment earlier! 

Here's the Claude distillation based on my reasoning on why to use it:

Reclaim is useful because it lets you assign different priorities to tasks and meetings, automatically scheduling recurring meetings to fit your existing commitments while protecting time for important activities. 

For example, you can set exercising three times per week as a priority 3 task, which will override priority 2 meetings, ensuring those exercise timeblocks can't be scheduled over. It also automatically books recurrent meetin... (read more)

Thanks Jacques! I was looking for an upgrade to some of my LLM tools. I was looking for some IDEs and I'll check that out.

The only tip I've got is using reclaim.ai instead of calendly for automatic meeting scheduling, it slaps.

2
Neel Nanda
What does reclaim give you? I've never heard of it, and the website is fairly uninformative

Thanks! That post adresses what I was pointing at a lot better than I did in mine. 

I can see from your response that I didn't get across my point as well as I wanted to but I appreciate the answer none the less!

It was more a question of what leads to the better long-term consequences rather than combining them.

It seems plausible animals have moral patienthood and so the scale of the problem is larger for animals whilst also having higher tractability. At the same time, you have cascading effects of economic development into better decision making. As a longtermist, this makes me very uncertain on where to focus resources. I will therefore put myself centrally to signal my high uncertainty.

I think that still makes sense under my model of a younger and less tractable field? 

Experience comes partly from the field being viable for a longer period of time since there can be a lot more people who have worked in that area in the past. 

The well-described steps and concrete near-term goals can be described as a lack of easy tractability? 

I'm not saying that it isn't the case that the proposals in longtermism are worse today but rather that it will probably look different in 10 years? A question that pops up for me is about how great t... (read more)

2
Arepo
That might be part of the effect, but I would think it would apply more strongly to EA community building than AI (which has been around for several decades with vastly more money flowing into it) - and the community projects were maybe slightly better over all? At least not substantially worse. I don't really buy that concrete steps are hard to come up with for good AI or even general longtermism projects - one could for e.g. aim to show or disprove some proposition in a research program, aim to reach some number of views, aim to produce x media every y days (which IIRC one project did), or write x-thousand words or interview x industry experts, or use some tool for some effect, or any one of countless ways of just breaking down what your physical interactions will be with the world between now and your envisioned success. 

I enjoyed the post and I thought the platform for collective action looked quite cool.

I also want to mention that I think tractability is just generally a really hard thing for longtermism. It's also a newer field and so on expectation I think you should just believe that the projects will look worse than in animal welfare. I don't think there's any need for psychoanalysis of the people in the space even though it has its fair share of wackos.

4
Arepo
Fwiw it felt like a more concrete difference than that. My overall sense is that the animal welfare projects tended to be backed by multiple people with years of experience doing something relevant, have a concrete near term goal or set of milestones, and a set of well-described steps for moving forwards, while the longtermist/AI stuff tended to lack some or all of that.

Great point, I did not think of the specific claim of 5% when thinking of the scale but rather whether more effort should be spent in general.

My brain basically did a motte and baily on me emotionally when it comes to this question so I appreciate you pointing that out!

It also seems like you're mostly critiquing the tractability of the claim and not the underlying scale nor neglectedness?

It kind of gives me some GPR vibes as for why it's useful to do right now and that dependent on initial results either less or more resources should be spent?

3
JWS 🔸
Yep, everyone agrees it's neglected. My strongest critique is the tractability, which may be so low as to discount astronomical value. I do take a lot of issue with the scale as well though. I think that needs to be argued for rather than assumed. I also think trade-offs from other causes need to be taken into account at some point too. And again, I don't think there's no arguments that can make traction on the scale/tractability that can make AI Welfare look like a valuable cause, but these arguments clearly weren't made (imho) in AWDW

Super exciting! 

I just wanted to share a random perspective here: Would it be useful to model sentience alongside consciousness itself? 

If you read Daniel Dennett's book Kinds of Minds or take some of the Integrated Information Theory stuff seriously, you will arrive at this view of a field of consciousness. This view is similar to Philip Goff's or to more Eastern traditions such as Buddhism. 

Also, even in theories like Global Workspace Theory, the amount of localised information at a point in time matters alongside the type of information p... (read more)

Wild animal welfare and longtermist animal welfare versus farmed animal welfare? 

There's this idea of the truth as an asymmetric weapon; I guess my point isn't necessarily that the approach vector will be something like:
Expert discussion -> Policy change

but rather something like
Experts discussion -> Public opinion change -> Policy Change

You could say something about memetics and that it is the most understandable memes that get passed down rather than the truth, which is, to some extent, fair. I guess I'm a believer that the world can be updated based on expert opinion. 

For example, I've noticed a trend in the AI Safety d... (read more)

3
Toby Tremlett🔹
I think this is a good description of the kind of scepticism I'm attracted to, perhaps to an irrational degree. Thanks for describing it! I like your point about AI Safety. It seems at least a bit true.  I'll update my vote on the banner to be a bit less sceptical- I think my scepticism of the potential for us to know whether AI is conscious is a major part of my disagreement with the debate statement. I don't endorse the level of scepticism I hold. Thanks!

Yeah, I guess the crux here is to what extent we actually need public support or at least what type of public support that we need for it to become legislation?

If we can convince 80-90% of the experts, then I believe that this has cascading effects on the population, and it isn't like AI being conscious is something that is impossible to believe either. 
I'm sure millions of students have had discussions about AI sentience for fun, and so it isn't like fully out of the Overton window either.

I'm curious to know if you disagree with the above or if there is another reason why you think research won't cascade to public opinion? Any examples you could point towards? 

2
Toby Tremlett🔹
I don't have an example to mind exactly, but I'd expect you could find one in animal welfare. Where there are agricultural interests pushing against a decision, you need a public campaign to counter them. We don't live in technocracies-- representatives need to be shown that there is a commensurate interest in favour of the animals. On less important issues/ legislation which can be symbolic but isn't expected to be used- experts can have a more of a role. I'd expect that the former category is the more important one for digital minds. Does that make sense? I'm aware its a bit too stark of a dichotomy to be true. 

A crux that I have here is that research that takes a while to explain is not going to inspire a popular movement. 

Okay, what comes to mind for me here is quantum mechanics and how we've come up with some pretty good analogies to explain parts of it. 

Do we really need to communicate the full intricacies of AI sentience to say that an AI is conscious? I guess that this isn't the case.

The world where EA research and advocacy for AI welfare is most crucial is one where the reasons to think that AI systems are conscious are non-obvious, such that we

... (read more)
3
Toby Tremlett🔹
Thanks! I'm also excited about this week- it's really cool to see how many people have already voted- goes well beyond my expectations. 
3
Toby Tremlett🔹
I think this is a great point, and might change my mind. However, if these consciousness evals become burdensome for AI companies, I would imagine we would need a public push in support of them in order for them to be enforced, especially through legislation. Then we get back to my dichotomy, where if people think AI is obviously conscious (whether or not it is) we might get legislation, and if they don't, I can only imagine some companies doing it half-heartedly/ voluntarily until it becomes too costly (as is, arguably, the current state of safety evals). 

Damn, I really resonated with this post. 

I share most of your concerns, but I also feel that I have some even more weird thoughts on specific things, and I often feel like, "What the fuck did I get myself into?"

Now, as I've basically been into AI Safety for the last 4 years, I've really tried to dive deep into the nature of agency. You get into some very weird parts of trying to computationally define the boundary between an agent and the things surrounding it and the division between individual and collective intelligence just starts to break down a ... (read more)

So I've been working in a very adjacent space to these ideas for the last 6 months and I think that the biggest problems that I have with this is just the feasibility of it.

That being said we have thought about some ways of approaching a GTM for a very similar system. Thr system I'm talking about here is an algorithm to improve interpretability and epistemics of organizations using AI.

One is to sell it as a way to "align" management teams lower down in the organization for the C-suite level since this actually incentivises people to buy it.

A second one is ... (read more)

1
Mike P. Sinn
Hi, Jonas!  Do you have a link to more info about your project?
2
Ozzie Gooen
That sounds really neat, thanks for sharing!

I appreciate you putting out a support post of someone who might have some EA leanings that would be good to pick up on. I may or may not have done so in the past and then removed the post because people absolutely shat on it on the forum 😅 so respect.

2
PeterSlattery
Thanks, Jonas! I appreciate the support :P 

I guess I felt that a lot of the post was arguing under a frame of utilitarianism which is generally fair I think. When it comes to "not leaving a footprint on the future" what I'm referring to is epistemic humility about the correct moral theories. I'm quite uncertain myself about what is correct when it comes to morality with extra weight on utilitarianism. From this, we should be worried about being wrong and therefore try our best to not lock in whatever we're currently thinking. (The classic example being if we did this 200 years ago we might still ha... (read more)

Load more