Hide table of contents

TL;DR

  • We hosted an AI Safety “Thinkathon” in Chile. We had participation from 40 students with differing skill levels and backgrounds, with groups totalling 13 submissions. 
  • We see potential in:
    • Similar introductory events aiming for a broad audience
    • Collaborating more often with student organizations
    • Leveraging remote help from external mentors
  • We experimented with an alternative naming, having remote mentors, different problem sources, and student organization partnerships, with varying results.
  • We could have improved planning and communicating the difficulty of challenges.

Introduction

In February, we ran the first AI Safety Hackathon in Chile (and possibly in all of South America[1]). This post provides some details about the event, a teaser of some resulting projects and our learnings throughout.

A team working on their proposal during the second day.

Goals and overview of the event

The hackathon was meant to kick-start our nascent AI Safety Group at UC Chile, generating interest in AI Safety and encouraging people to register for our AGI Safety Fundamentals course group.

It ran between the 25th and the 28th of February, the first two days being in-person events and the other two serving as additional time for participants to work on their proposals, with some remote assistance on our part. Participants formed teams of up to four people, and could choose to assist either virtually (through Discord) or in-person (on the first two days).

We had help from Apart Research and partial funding from AI Alignment Awards.

Things we experimented with

  • Aiming for a broad audience, we named the event “Thinkathon” (instead of hackathon) and provided plenty of introductory material alongside the proposed problems. 
    • We think this was the right choice, as the desired effect was reflected on the participant demographics (see below).
    • We could have been better at preparing participants. Some participants suggested we could have done an introductory workshop.
  • We incorporated the two problems from the AI Alignment Awards (Goal Misgeneralization and the Shutdown problem), alongside easier, self-contained problems aimed at students with different backgrounds (like policy or psychology). 
    • We think most teams weren't prepared to tackle the AI Alignment Awards challenges. Most teams (77%) chose them initially regardless of their experience, getting stuck quickly.
    • This might have worked better by communicating difficulty more clearly, as well as emphasizing that aiming for incremental progress rather than a complete solution is a better strategy for a beginner's hackathon.
  • As we don't know many people with previous experience in AIS in Chile, we got help from external mentors, which connected remotely to help participants.
    • We think this was a good decision, as participants rated mentor support highly (see below).
  • We collaborated actively with two student governments from our university (the Administration and Economics Student Council and the Engineering Student Council). They helped with funding, logistics and outreach.
    • We think this was an excellent choice, as they provided a much broader platform for outreach and crucial logistics help.
    • We had a great time working with them, and they were eager to work with us again!

Things that went well

  • 40 people attended in-person and 10 people remotely (through Discord), we were surprised by both the high number of attendants and the preference[2] for in-person participation.
  • We had a total of 13 submitted proposals, much higher than expected.
  • While all proposals were incremental contributions, most were of high quality.
  • Skill level and majors varied significantly, going from relatively advanced CS students to freshmen from other fields (like economics). We were aiming for diversity, so this is a win.
  • While we hugely delayed starting the AGISF program (we opened applications very recently), many participants were eager to apply to AGISF.

Things we could have done better

(Also check out the “Things we experimented with” section)

  • We had some issues navigating funding. This was in good part because we failed to anticipate some costs. In the end, we ended up spending around $20 USD per person total on food and catering. The collaborating student organizations covered most of the funding for prizes (around $15 USD / pp).
  • Logistics should have been planned with significantly more anticipation. We were overconfident about being able to sort out some details last minute. It turned out our university couldn't host the event (due to the academic calendar), and we were forced to find two new venues in a couple of days (which we were able to get for free).
  • At the end of the last day of the in-person phase, we didn't do a proper wrap-up. This contributed to some teams feeling a bit lost during the remote phase.
  • We didn't do an impressive job at positioning the AI Safety group, as we could have talked more about it during the event, and be quicker to follow-up on participants' interest after the event.

Challenge problems

We presented participants with the following six challenges. The linked write-ups are in Spanish, as they were made for participants.

Highlighted submissions

Some highlighted proposals from the hackathon:

Investigating the Relationship Between Priming with Multiple Traits and Language Model Truthfulness in GPT-3: A research project exploring how different traits placed in prompts influence the truthfulness of GPT-3 based models, following this idea from AISI

The challenge of Goal Misgeneralization in AI: Causes and Solutions: A general literature review on goal misgeneralization that also suggests a mix of solutions inspired heavily by previous work. 

All proposals (some in Spanish) can be seen here.

Attendance Demographics

  • Around 50% of students were majoring in engineering (mostly CS engineering[7]), 21% in a (new) purely CS major, 11% in business and economics, and the rest were from a mix of majors (neuroscience, maths, biochemistry, medicine).
  • About 75% of participants were from UC Chile, our university, while about 25% were from other universities.
  • There was a significant gender imbalance (only 20% were women or non-binary people). We don't have great ideas on how to improve this.
  • To our surprise, around 54% of participants were incoming freshmen.

Insights from participants

Anecdotally:

  • It was surprisingly easy to talk about concerns related to AI Safety with everyone, including people with no previous exposure to them.
  • Most people didn't have trouble with understanding the problems themselves, independent of their background (except perhaps for goal misgeneralization).

We also ran a quick after-event survey (n=26):

  • Participants gave a 77% Net Promoter Score (4.6/5 likelihood to recommend), indicating high satisfaction with the event.
  • Frequent negative feedback included lack of food on the second day (n=5), lack of activities to socialize or meet other participants (n=4) and that the venues used were too far away (n=4).
  • 76% hadn't heard of AI Safety before the event.[8]
  • Interest in working or doing research in AI Safety and AI more generally increased significantly, with a moderate effect size for the first (r = 0.54) and a small effect size (r = 0.32) for the latter.[9] 
  • Participants reported having learned significantly about both AI in general and AI Safety (4.0/5 and 3.8/5 respectively) and valued mentor help highly (4.4/5)

Also, some random pictures:

Acknowledgements

We want to thank the following people for their amazing help:

  1. ^

    We might or might have not chosen to write South America instead of Latin America to conveniently exclude Mexico

  2. ^

    Especially considering that participating in-person meant showing up at 9:00  of a summer vacation day in an inconveniently-located venue.

  3. ^
  1. ^

    Adapted from this idea by Caroline Jeanmaire. 

  2. ^

     Taken from this idea by @Sabrina Zaki, Luke Ring and Aleks Baskakovs, that in turn comes from their (winning) proposal for Apart Research's language model hackathon.

  3. ^

     Inspired the ideas in @Geoffrey Irving and Amanda Askell's AI Safety Needs Social Scientists and Riccardo Volpato's Research ideas to study humans with AI Safety in Mind.

  4. ^

    People from CS Engineering take common courses in engineering, but are otherwise similar to CS.

  5. ^

    This was explicitly defined in contrast to Ethics of AI or similar fields.

  6. ^

    A Wilcoxon signed-rank test showed a statistically significant increase in interest in AI (W = 115.5, p < 0.001, r = 0.32) as well as AI Safety (W = 190.0, p < 0.001, r = 0.54). This was based on retrospective Likert-style questions on interest. It was hard to account for survey effects. 

Show all footnotes
No comments on this post yet.
Be the first to respond.
Curated and popular this week
Sam Anschell
 ·  · 6m read
 · 
*Disclaimer* I am writing this post in a personal capacity; the opinions I express are my own and do not represent my employer. I think that more people and orgs (especially nonprofits) should consider negotiating the cost of sizable expenses. In my experience, there is usually nothing to lose by respectfully asking to pay less, and doing so can sometimes save thousands or tens of thousands of dollars per hour. This is because negotiating doesn’t take very much time[1], savings can persist across multiple years, and counterparties can be surprisingly generous with discounts. Here are a few examples of expenses that may be negotiable: For organizations * Software or news subscriptions * Of 35 corporate software and news providers I’ve negotiated with, 30 have been willing to provide discounts. These discounts range from 10% to 80%, with an average of around 40%. * Leases * A friend was able to negotiate a 22% reduction in the price per square foot on a corporate lease and secured a couple months of free rent. This led to >$480,000 in savings for their nonprofit. Other negotiable parameters include: * Square footage counted towards rent costs * Lease length * A tenant improvement allowance * Certain physical goods (e.g., smart TVs) * Buying in bulk can be a great lever for negotiating smaller items like covid tests, and can reduce costs by 50% or more. * Event/retreat venues (both venue price and smaller items like food and AV) * Hotel blocks * A quick email with the rates of comparable but more affordable hotel blocks can often save ~10%. * Professional service contracts with large for-profit firms (e.g., IT contracts, office internet coverage) * Insurance premiums (though I am less confident that this is negotiable) For many products and services, a nonprofit can qualify for a discount simply by providing their IRS determination letter or getting verified on platforms like TechSoup. In my experience, most vendors and companies
 ·  · 4m read
 · 
Forethought[1] is a new AI macrostrategy research group cofounded by Max Dalton, Will MacAskill, Tom Davidson, and Amrit Sidhu-Brar. We are trying to figure out how to navigate the (potentially rapid) transition to a world with superintelligent AI systems. We aim to tackle the most important questions we can find, unrestricted by the current Overton window. More details on our website. Why we exist We think that AGI might come soon (say, modal timelines to mostly-automated AI R&D in the next 2-8 years), and might significantly accelerate technological progress, leading to many different challenges. We don’t yet have a good understanding of what this change might look like or how to navigate it. Society is not prepared. Moreover, we want the world to not just avoid catastrophe: we want to reach a really great future. We think about what this might be like (incorporating moral uncertainty), and what we can do, now, to build towards a good future. Like all projects, this started out with a plethora of Google docs. We ran a series of seminars to explore the ideas further, and that cascaded into an organization. This area of work feels to us like the early days of EA: we’re exploring unusual, neglected ideas, and finding research progress surprisingly tractable. And while we start out with (literally) galaxy-brained schemes, they often ground out into fairly specific and concrete ideas about what should happen next. Of course, we’re bringing principles like scope sensitivity, impartiality, etc to our thinking, and we think that these issues urgently need more morally dedicated and thoughtful people working on them. Research Research agendas We are currently pursuing the following perspectives: * Preparing for the intelligence explosion: If AI drives explosive growth there will be an enormous number of challenges we have to face. In addition to misalignment risk and biorisk, this potentially includes: how to govern the development of new weapons of mass destr
Dr Kassim
 ·  · 4m read
 · 
Hey everyone, I’ve been going through the EA Introductory Program, and I have to admit some of these ideas make sense, but others leave me with more questions than answers. I’m trying to wrap my head around certain core EA principles, and the more I think about them, the more I wonder: Am I misunderstanding, or are there blind spots in EA’s approach? I’d really love to hear what others think. Maybe you can help me clarify some of my doubts. Or maybe you share the same reservations? Let’s talk. Cause Prioritization. Does It Ignore Political and Social Reality? EA focuses on doing the most good per dollar, which makes sense in theory. But does it hold up when you apply it to real world contexts especially in countries like Uganda? Take malaria prevention. It’s a top EA cause because it’s highly cost effective $5,000 can save a life through bed nets (GiveWell, 2023). But what happens when government corruption or instability disrupts these programs? The Global Fund scandal in Uganda saw $1.6 million in malaria aid mismanaged (Global Fund Audit Report, 2016). If money isn’t reaching the people it’s meant to help, is it really the best use of resources? And what about leadership changes? Policies shift unpredictably here. A national animal welfare initiative I supported lost momentum when political priorities changed. How does EA factor in these uncertainties when prioritizing causes? It feels like EA assumes a stable world where money always achieves the intended impact. But what if that’s not the world we live in? Long termism. A Luxury When the Present Is in Crisis? I get why long termists argue that future people matter. But should we really prioritize them over people suffering today? Long termism tells us that existential risks like AI could wipe out trillions of future lives. But in Uganda, we’re losing lives now—1,500+ die from rabies annually (WHO, 2021), and 41% of children suffer from stunting due to malnutrition (UNICEF, 2022). These are preventable d
Relevant opportunities