2 min read 12

7

Labor unions are associations of workers that negotiate with employers. A union for AI workers such as data scientists, hardware and software engineers could organise labor to counterbalance the influence of shareholders or political masters.

Importantly, unions could play a unique, direct role in redirecting or slowing down the rapid development of AI technology across multiple companies when there is a concern about safety and race dynamics. With difficult to replace expertise, they could do so independent of employers wishes.

Unions often negotiate with multiple companies simultaneously, including in industries where competition is fierce. By uniting workers across AI labs, unions could exert significant collective bargaining power to demand a pause or slower, more cautious development of AI systems with a strong emphasis on safety.

If union demands are not met, they can and in history have organised workplace slowdowns or work stoppages as a form of protest or negotiation tactic. If workers across various AI companies and countries organise together, they can coordinate slowdowns or strikes that affect multiple companies simultaneously.

If the AI safety community seeded or nurtured an AI workers union they could also help embed a longtermist culture of safety. Unions already have a proven track record of prioritising and achieving safety in various fields more effectively than employers alone. They often foster a culture of safety that encourages workers to be proactive in identifying and addressing safety concerns. Unions also often provide protection and support for employees who report safety violations or concerns. This encourages workers to come forward without fear of retaliation, ensuring that safety issues are addressed promptly.

With roots in the AI safety community, an AI workers union could advocate for AI safety in government and corporate policies and regulations with greater independence from profit-motives.

Some practical considerations and open questions:

Google tells me there are already some unions for data scientists and software engineers. However their relevance relative to the scale of the challenge is negligible. That is not to say an AI workers union is not feasible. Support for unions in the United States has risen from 65% before the pandemic to 71% in 2022, the highest support level since 1965. Whether or not that is reflective of the tech industry I cannot say.

If some countries unionise AI workers more readily than others what will the geopolitical considerations be? More harmful than good? Will restrictions on the activities of union in different countries affect the efficacy of union organising for AI safety.

Since AI workers are relatively well remunerated (what Marxists would call petty bourgeois) they may lack the class conscience to unionise. On the other hand, these workers will be well placed to contribute funding for a union to scale and punch above its weight in members. Could a critical mass of AI workers be recruited to collectively bargain effectively?

Defining the union's scope can enhance its influence and bargaining power but requires careful planning. The choice of occupations to incorporate in the union might include data scientists, machine learning engineers, hardware experts. But other workers are involved in AI-related work such as ethicists at a university, sailors shipping semiconductor by sea or policy professionals at an AI lab. Should they be incorporated?

Should a union of AI workers be its own entity, entities or part of a multipurpose union like the Industrial Workers of the World (IWW)? Should AI safety activists nurture existing data science or software unions or start their own initiatives? Do AI workers share common concerns that are distinct from those of workers in other industries?

I don't know, but these questions are possible directions those reading may want to explore and comment on.

Comments12


Sorted by Click to highlight new comments since:

A union for AI workers such as data scientists, hardware and software engineers could organise labor to counterbalance the influence of shareholders or political masters.

 

It's not obvious to me that AI workers would want a more cautious approach than AI shareholders, AI bosses, and so on. Whether or not this would be the case seems to me to be the main crux behind whether this would be net positive or net harmful.

Even if they were slightly more cautious than management, if they were less cautious than policymakers it could still be net negative due to unions' lobbying abilities.

Granted, in principle you could also have a situation where they're less cautious than management but more cautious than policymakers and it winds up being net positive, though I think that situation is pretty unlikely. Agree the consideration you raised is worth paying attention to.

I had explicitly considered this in drafting and whether to state that crux. If so, it could be an empirical question of whether there is greater support from the workers or management, or receptiveness to change.

I did not because I now think the question is not whether AI workers are more cautious than AI shareholders, but whether AI firms where unionised AI workers negotiate with AI shareholders would be more cautious. To answer that question, I think so

Edit: to summarise, the question is not whether unions (in isolation) would be more cautious, but whether an system of management (and policymakers) bargaining with a union would be more cautious - and yes it probably would

I've thought about this before and talked to a couple people in labs about it. I'm pretty uncertain if it would actually be positive. It seems possible that most ML researchers and engineers might want AI development to go as quickly or more than leadership if they're excited about working on cutting edge technologies or changing the world or for equity reasons. I remember some articles about how people left Google for companies like OpenAI because they thought Google was too slow, cautious, and lost its "move fast and break things" ethos. 

As you have said there are examples of individuals have left firms because they feel their company is too cautious. Conversely there are individuals who have left for companies that priorities AI safety.

If we zoom out and take the outside view, it is common for those individuals who form a union to take action to slow down or stop their work or take action to improve safety. I do not know an example of a union that has instead prioritised acceleration.

That's a good point. Although 1) if people leave a company to go to one that prioritizes AI safety, then this means there are fewer workers at all the other companies who feel as strongly. So a union is less likely to improve safety there. 2) It's common for workers to take action to improve safety conditions for them, and much less common for them to take action on issues that don't directly affect their work, such as air pollution or carbon pollution, and 3) if safety inclined people become tagged as wanting to just generally slow down the company, then hiring teams will likely start filtering out many of the most safety minded people. 

Thanks for writing this; I've thought about this before, it seems like an under-explored (or under-exploited?) idea. 

Another point: even if ML engineers, software devs etc either could not be persuaded to unionize, or would accelerate AI development if they could, maybe other labour unions could still exert pressure. E.g., workers in the compute or hardware supply chain; HR, cleaners, ops, and other non-technical staff who work at AI companies? Perhaps strong labour unions in sectors that are NOT obviously related to AI could be powerful here, e.g. by consumer boycotts (e.g., what if education union members committed to not spending money on AI products unless and until the companies producing them complied with certain safety measures?)

Some recent polls suggest that the idea of slowing down AI is already popular among US citizens (72% want to slow it down). My loose impressions are also that (i) most union members and organizers are on the political left (ii) many on the left are already sceptical about AI, for reasons related to (un)employment, plagiarism (i.e. critics of art AI's use of existing art), capitalism (tech too controlled by powerful interests), algorithmic bias. So this might not be an impossible sell, if AI safety advocates communicate about it in the right way.

To your first para - yes I wonder how unionised countries and relevant sectors are in bottlenecks in the compute supply chain - Netherlands, Japan and Taiwan. I don't know enough about the efficacy of boycotts to comment on the union led boycotts idea.

I've raised this in response to another comment but I want to also address here the concern that workers who join a union would organise to accelerate the development of AI. I think that is very unlikely - the history of unions is a strong tradition of safety, slowing down or stopping work. I do not know an example of a union that has instead prioritised acceleration but there's probably some and it would get grey as you move into the workers self-management space.

Yeah I don't have a strong opinion about whether they would accelerate it - I was just saying, even if some workers would support acceleration, other workers could work to slow it down.

One reason that developers might oppose slowing down AI is that it would put them out of work, wouldn't it? (Or threaten to). So if someone is not convinced that AI poses a big risk, or thinks that pausing isn't the best way to address the risk, then lobbying to slow down AI development would be a big cost for no obvious benefit. 

Something feels off about this Article. It is not really discussed what the AI workers could want or believe, or how to convince them that slowing down AI would delay or aviod extinction of humanity.

Are you assuming a world where the risk of extinction from AGI is widely accepted among AI workers? (In this case, why are they still working on the thing that potentially kills everyone?) If the workers do not believe in (large) risks of extinction from AI, how do you want to recruit them into your union? This seems hard if you want to be honest about the main goal of the union?

I don't think this is predicated on those assumptions.

My assumptions are:

  • AI workers who join a union are more likely to care about safety than AI workers who do not join a union. That is because the history of unions suggests that unions promote a culture of safety

  • Unionised AI workers will be more organised in influencing their workplace than non unionised AI workers. That is because of their ability to co-ordinate collectively

Therefore:

  • Unionisation of AI workers would encourage a culture of safety

Furthermore, these unions could be in a position to implement AI safety policies.

Curated and popular this week
 ·  · 17m read
 · 
TL;DR Exactly one year after receiving our seed funding upon completion of the Charity Entrepreneurship program, we (Miri and Evan) look back on our first year of operations, discuss our plans for the future, and launch our fundraising for our Year 2 budget. Family Planning could be one of the most cost-effective public health interventions available. Reducing unintended pregnancies lowers maternal mortality, decreases rates of unsafe abortions, and reduces maternal morbidity. Increasing the interval between births lowers under-five mortality. Allowing women to control their reproductive health leads to improved education and a significant increase in their income. Many excellent organisations have laid out the case for Family Planning, most recently GiveWell.[1] In many low and middle income countries, many women who want to delay or prevent their next pregnancy can not access contraceptives due to poor supply chains and high costs. Access to Medicines Initiative (AMI) was incubated by Ambitious Impact’s Charity Entrepreneurship Incubation Program in 2024 with the goal of increasing the availability of contraceptives and other essential medicines.[2] The Problem Maternal mortality is a serious problem in Nigeria. Globally, almost 28.5% of all maternal deaths occur in Nigeria. This is driven by Nigeria’s staggeringly high maternal mortality rate of 1,047 deaths per 100,000 live births, the third highest in the world. To illustrate the magnitude, for the U.K., this number is 8 deaths per 100,000 live births.   While there are many contributing factors, 29% of pregnancies in Nigeria are unintended. 6 out of 10 women of reproductive age in Nigeria have an unmet need for contraception, and fulfilling these needs would likely prevent almost 11,000 maternal deaths per year. Additionally, the Guttmacher Institute estimates that every dollar spent on contraceptive services beyond the current level would reduce the cost of pregnancy-related and newborn care by three do
 ·  · 2m read
 · 
I speak to many entrepreneurial people trying to do a large amount of good by starting a nonprofit organisation. I think this is often an error for four main reasons. 1. Scalability 2. Capital counterfactuals 3. Standards 4. Learning potential 5. Earning to give potential These arguments are most applicable to starting high-growth organisations, such as startups.[1] Scalability There is a lot of capital available for startups, and established mechanisms exist to continue raising funds if the ROI appears high. It seems extremely difficult to operate a nonprofit with a budget of more than $30M per year (e.g., with approximately 150 people), but this is not particularly unusual for for-profit organisations. Capital Counterfactuals I generally believe that value-aligned funders are spending their money reasonably well, while for-profit investors are spending theirs extremely poorly (on altruistic grounds). If you can redirect that funding towards high-altruism value work, you could potentially create a much larger delta between your use of funding and the counterfactual of someone else receiving those funds. You also won’t be reliant on constantly convincing donors to give you money, once you’re generating revenue. Standards Nonprofits have significantly weaker feedback mechanisms compared to for-profits. They are often difficult to evaluate and lack a natural kill function. Few people are going to complain that you provided bad service when it didn’t cost them anything. Most nonprofits are not very ambitious, despite having large moral ambitions. It’s challenging to find talented people willing to accept a substantial pay cut to work with you. For-profits are considerably more likely to create something that people actually want. Learning Potential Most people should be trying to put themselves in a better position to do useful work later on. People often report learning a great deal from working at high-growth companies, building interesting connection
 ·  · 1m read
 · 
Need help planning your career? Probably Good’s 1-1 advising service is back! After refining our approach and expanding our capacity, we’re excited to once again offer personal advising sessions to help people figure out how to build careers that are good for them and for the world. Our advising is open to people at all career stages who want to have a positive impact across a range of cause areas—whether you're early in your career, looking to make a transition, or facing uncertainty about your next steps. Some applicants come in with specific plans they want feedback on, while others are just beginning to explore what impactful careers could look like for them. Either way, we aim to provide useful guidance tailored to your situation. Learn more about our advising program and apply here. Also, if you know someone who might benefit from an advising call, we’d really appreciate you passing this along. Looking forward to hearing from those interested. Feel free to get in touch if you have any questions. Finally, we wanted to say a big thank you to 80,000 Hours for their help! The input that they gave us, both now and earlier in the process, was instrumental in shaping what our advising program will look like, and we really appreciate their support.