Hide table of contents

Recently my team and I have been working on FindingConsensus.AI. It’s a site to show how public figures would answer specific questions about Artificial Intelligence (but it could work for anything). Often when thinking, I defer to figures I trust, so seeing a spread of public figures gives me a head start. I hope that if we all knew what we all thought, we’d be better at finding consensus, too.

How does the site work?

The questions are currently focused around a Californian AI bill, but this article isn’t deeply focused on that. 

 

Each question has a title and context, and displays public figures and organisations in their answers to a question. 

Above, the question is about how many companies the bill, SB 1047, will affect (in the language of the bill, these are called “covered models”). This seems like a pretty important empirical question. Who will the bill affect? If we can’t even agree on that, what can we agree on?

The figure shows estimates of opinions of different public figures and organisations. My team and I read statements from them and guessed their answers to these questions. 

As you can see, even on this relatively clear-cut question, there is a lot of disagreement.

The full set of questions (at time of posting) are shown at the bottom of this blog. 

Wait, you just guess their opinions?

As regular readers know, I am a staunch defender of guessing. Have I gone too far this time? 

It can be hard to find a quote that directly answers a question, even when a person’s opinion feels fairly clear. Public figures hold opinions about AI, many of them tweet about it all day. We feel like we know their views on a topic. But often this isn’t because they ever commented explicitly and directly on a specific question in a couple of sentences. 

I don’t think this should stop us from showing their views. In the case of say a16z (a Venture Capital firm) and SB 1047 (this bill), they have made their views clear - they think it will damage the ecosystem and perhaps even drive companies out of San Francisco. It is reasonable to represent this on the chart as them saying it will almost certainly hamstring the AI industry, even without perfect quotes. 

I am tempted to guess the opinions of organisations even where there are no quotes at all, but I sense this would go too far. The upside would be that currently it’s beneficial to be ambiguous or cagey on specifics even when supporting or speaking against the bill. For those vague positions, I could anchor them to some kind of base rate and guess their opinions entirely. However, I think this would surprise readers and legitimacy feels important to this site and that people don’t expect entirely guessed positions. 

There are risks to my current estimation strategy, notably that I’m wrong. It would be bad for people to see a misrepresentation of someone’s views because of FindingConsensus.AI. I’m trying to reduce this in two ways. First, there is a way for people to correct quotes about them by getting in touch with me. Second I attempt to make clear that these are estimates. If some journalist uses the estimates without conveying that, I think that’s their error. But still miscommunication is something I’m wary of. 

Is this valuable?

I don’t think it’s clearly valuable yet. It’s not like it gets 1000s of daily hits. But I hope it can create value in a few ways:

Help see where people disagree. I often find discussion on important topics repetitive. In the discussion of this bill some people love SB 1047 and some people hate it. But why? Well having looked into it, I have some guesses. People disagree on if it will:

  • Reduce large risks from AI
  • Hamstring the AI industry
  • ‘Cover’ many AI models
  • De-facto ban open source

I wish there was a site like this for all the issues I discuss regularly.

Point to places where one can dig deeper. When I find a disagreement, it hints that people have different models of the world. I can ask questions to surface these. Let’s look at one question: The number of models the bill will cover in the next 5 years. It’s a relatively well-defined number that will be known in 2029. And yet look at the breadth of answers:

 

What is going on here? Why are some people so confident that it won’t affect more than 10 companies and others are so confident that it will. To me this suggests a deeper disagreement.

Dean Ball, a Research Fellow at George Mason University, gives his reasons (emphasis mine): 

SB 1047 will almost certainly cover models from more than 10 companies in 2029. ... the bill is not just about language models ... [also] we really have no idea how the bill's "training cost" threshold will be calculated in practice 

Part of this is that he is unsure how training costs will be measured. If we wanted to, we could drill down on this point, trying to get opinions on whether the definition of training costs will be taken restrictively or expansively. 

I expect that all these disagreements are because of very different models of the world. 

Understand the discussion better. If I understand people’s models of the world I can change their mind (or mine) and do deals with them. I softly support the bill (SB 1047), but having done this work I find it a lot easier to empathise with some who don’t like it (others are, in my opinion, arguing in bad faith). To give the views I empathise with. much regulation is poor, and too rarely repealed. If I didn’t think the risks were so high I wouldn’t want a fledgeling industry to be heavily regulated either. 

Supporting journalists. I’ve seen 1 journalist say this kind of website is useful to them, providing quotes they can use in articles. I can imagine a better version would be more useful still. 

How could this improve discussions?

 

 

I am my father’s son. And so I have drawn you some diagrams. Let’s start on the left. 

I find many discussions are like the film 50 First Dates. Adam Sandler dates an amnesiac, Drew Barrymore, slowly improving his dating strategy each day while she forgets. To be clear, we’re the Drew Barrymore in this analogy. We have the same discussions on AI, trans stuff, politics and forget any progress. Discussions get little better over time. Some people leave, others become angry or less epistemically rigorous.

Now on the right:

I hope we can move to a different mode, perhaps like checkpoints in a game like Crash Bandicoot. You have part of a discussion, you save the progress of the discussion, you create common knowledge of that saved progress. When you start again, you take this common knowledge into account. Over time, the discussion leaves out bits that everyone agrees aren’t important and focuses on the points of contention. Hopefully in time, there are deals and solutions. 

The discussion of this California bill has not been great, but the bill process itself has been pretty good, in my opinion:

  • The bill has gone through multiple rounds of rewrites
  • These have removed several things opponents hated but left most things that supporters like. Amongst other things, they increased the floor for the covered models, clarified open source rules and removed the government team it created. But liability for the most powerful models remains if companies cannot satisfy regulators.
  • These changes change likely won several powerful additional endorsements. I would guess the bill is broadly more able to pass than before these changes. 

I would like this process to happen organically. For people to say “I don’t like it if it’s like X but I would support Y” for specific and consensus values of Y. FindingConsensus.AI is part of that process.

Other parts of this process might be

  • Polling people to find out their positions to disseminate later
  • Using prediction markets to allow people to do deals across disagreements
  • Using a tool like viewpoints.xyz to find mutually agreeable solutions to disagreements

Learnings

I continue to think that building tools is good. This one doesn’t feel good enough yet, but I feel more confident in building. I would recommend to my past self to just build something he would use. If anyone else likes it, that’s gravy. 

People respond positively to the framing of the site. They understand where they fit into it and several were happy to tweet quotes specifically to go onto the site.

I spoke to people in different parties in this discussion who were interested in positive sum solutions. I don’t know how we get there but I think this is often underrated in AI discussions

Legitimacy is thorny. People were wary of me making too many editorial decisions to summarise a public figure. It is hard to find ways that a whole spread of people agree is fair.

Tell me what you think

Do you find it useful? What questions would you like to see public figure quotes on?

Thanks

Thanks to Rob Gordon, Ammiel Yawson, Josh Hart, Katja Grace for their work with me on this. Thanks to Danielle Fong, Dean Ball and Zvi Mowshowitz, Charles Foster and Alexander Campbell for their feedback and support.

This work was mainly funded by an organisation who wanted to use this tool for something else. They didn’t go forward so I repurposed some of the work. I don’t currently have permission to name them, but thanks for supporting this indirectly! 

Full screenshots

 

If you want to see all the quotes you’ll have to click through the panels on the site (FindingConsensus.AI). The starting quotes are randomised (and hence so are these screenshots).

 

 

 

 

 

Comments4


Sorted by Click to highlight new comments since:

I'd provide an option for figures to opt out of being included in this site.

Under your model would they be able to opt out of being written about in articles?

This feels incredibly useful in concept... If the goal is to build a sustainable system that you actually want to use, maybe create a library of figures to track, and use a language model to automatically sort statements as pertaining to certain issues and then gauge their stance with sentiment analysis... This could get pushed to you or a team member for approval before actually going on the site, and could save a lot of time.

Executive summary: The author describes FindingConsensus.AI, a new website that maps public figures' opinions on AI-related questions, aiming to improve discussions and find consensus on complex issues.

Key points:

  1. The site displays estimated opinions of public figures on specific questions, based on their statements and the author's interpretation.
  2. This approach helps identify areas of disagreement and deeper underlying models, potentially leading to more productive discussions.
  3. Challenges include ensuring accuracy of representations and maintaining legitimacy while making editorial decisions.
  4. The author hopes this tool can create "checkpoints" in discussions, allowing for progress and common knowledge to accumulate over time.
  5. Future improvements could include polling, prediction markets, and tools for finding mutually agreeable solutions.
  6. The project highlights the value of building practical tools and the potential for positive-sum solutions in AI discussions.

 

 

This comment was auto-generated by the EA Forum Team. Feel free to point out issues with this summary by replying to the comment, and contact us if you have feedback.

Curated and popular this week
 ·  · 23m read
 · 
Or on the types of prioritization, their strengths, pitfalls, and how EA should balance them   The cause prioritization landscape in EA is changing. Prominent groups have shut down, others have been founded, and everyone is trying to figure out how to prepare for AI. This is the first in a series of posts examining the state of cause prioritization and proposing strategies for moving forward.   Executive Summary * Performing prioritization work has been one of the main tasks, and arguably achievements, of EA. * We highlight three types of prioritization: Cause Prioritization, Within-Cause (Intervention) Prioritization, and Cross-Cause (Intervention) Prioritization. * We ask how much of EA prioritization work falls in each of these categories: * Our estimates suggest that, for the organizations we investigated, the current split is 89% within-cause work, 2% cross-cause, and 9% cause prioritization. * We then explore strengths and potential pitfalls of each level: * Cause prioritization offers a big-picture view for identifying pressing problems but can fail to capture the practical nuances that often determine real-world success. * Within-cause prioritization focuses on a narrower set of interventions with deeper more specialised analysis but risks missing higher-impact alternatives elsewhere. * Cross-cause prioritization broadens the scope to find synergies and the potential for greater impact, yet demands complex assumptions and compromises on measurement. * See the Summary Table below to view the considerations. * We encourage reflection and future work on what the best ways of prioritizing are and how EA should allocate resources between the three types. * With this in mind, we outline eight cruxes that sketch what factors could favor some types over others. * We also suggest some potential next steps aimed at refining our approach to prioritization by exploring variance, value of information, tractability, and the
 ·  · 5m read
 · 
[Cross-posted from my Substack here] If you spend time with people trying to change the world, you’ll come to an interesting conundrum: Various advocacy groups reference previous successful social movements as to why their chosen strategy is the most important one. Yet, these groups often follow wildly different strategies from each other to achieve social change. So, which one of them is right? The answer is all of them and none of them. This is because many people use research and historical movements to justify their pre-existing beliefs about how social change happens. Simply, you can find a case study to fit most plausible theories of how social change happens. For example, the groups might say: * Repeated nonviolent disruption is the key to social change, citing the Freedom Riders from the civil rights Movement or Act Up! from the gay rights movement. * Technological progress is what drives improvements in the human condition if you consider the development of the contraceptive pill funded by Katharine McCormick. * Organising and base-building is how change happens, as inspired by Ella Baker, the NAACP or Cesar Chavez from the United Workers Movement. * Insider advocacy is the real secret of social movements – look no further than how influential the Leadership Conference on Civil Rights was in passing the Civil Rights Acts of 1960 & 1964. * Democratic participation is the backbone of social change – just look at how Ireland lifted a ban on abortion via a Citizen’s Assembly. * And so on… To paint this picture, we can see this in action below: Source: Just Stop Oil which focuses on…civil resistance and disruption Source: The Civic Power Fund which focuses on… local organising What do we take away from all this? In my mind, a few key things: 1. Many different approaches have worked in changing the world so we should be humble and not assume we are doing The Most Important Thing 2. The case studies we focus on are likely confirmation bias, where
 ·  · 2m read
 · 
I speak to many entrepreneurial people trying to do a large amount of good by starting a nonprofit organisation. I think this is often an error for four main reasons. 1. Scalability 2. Capital counterfactuals 3. Standards 4. Learning potential 5. Earning to give potential These arguments are most applicable to starting high-growth organisations, such as startups.[1] Scalability There is a lot of capital available for startups, and established mechanisms exist to continue raising funds if the ROI appears high. It seems extremely difficult to operate a nonprofit with a budget of more than $30M per year (e.g., with approximately 150 people), but this is not particularly unusual for for-profit organisations. Capital Counterfactuals I generally believe that value-aligned funders are spending their money reasonably well, while for-profit investors are spending theirs extremely poorly (on altruistic grounds). If you can redirect that funding towards high-altruism value work, you could potentially create a much larger delta between your use of funding and the counterfactual of someone else receiving those funds. You also won’t be reliant on constantly convincing donors to give you money, once you’re generating revenue. Standards Nonprofits have significantly weaker feedback mechanisms compared to for-profits. They are often difficult to evaluate and lack a natural kill function. Few people are going to complain that you provided bad service when it didn’t cost them anything. Most nonprofits are not very ambitious, despite having large moral ambitions. It’s challenging to find talented people willing to accept a substantial pay cut to work with you. For-profits are considerably more likely to create something that people actually want. Learning Potential Most people should be trying to put themselves in a better position to do useful work later on. People often report learning a great deal from working at high-growth companies, building interesting connection