Chilean AIS Hackathon Retrospective

By Agustín Covarrubias 🔸, David Solar, Milan Weibel @ 2023-05-09T01:34 (+67)

TL;DR

Introduction

In February, we ran the first AI Safety Hackathon in Chile (and possibly in all of South America[1]). This post provides some details about the event, a teaser of some resulting projects and our learnings throughout.

A team working on their proposal during the second day.

Goals and overview of the event

The hackathon was meant to kick-start our nascent AI Safety Group at UC Chile, generating interest in AI Safety and encouraging people to register for our AGI Safety Fundamentals course group.

It ran between the 25th and the 28th of February, the first two days being in-person events and the other two serving as additional time for participants to work on their proposals, with some remote assistance on our part. Participants formed teams of up to four people, and could choose to assist either virtually (through Discord) or in-person (on the first two days).

We had help from Apart Research and partial funding from AI Alignment Awards.

Things we experimented with

Things that went well

Things we could have done better

(Also check out the “Things we experimented with” section)

Challenge problems

We presented participants with the following six challenges. The linked write-ups are in Spanish, as they were made for participants.

Highlighted submissions

Some highlighted proposals from the hackathon:

Investigating the Relationship Between Priming with Multiple Traits and Language Model Truthfulness in GPT-3: A research project exploring how different traits placed in prompts influence the truthfulness of GPT-3 based models, following this idea from AISI

The challenge of Goal Misgeneralization in AI: Causes and Solutions: A general literature review on goal misgeneralization that also suggests a mix of solutions inspired heavily by previous work. 

All proposals (some in Spanish) can be seen here.

Attendance Demographics

Insights from participants

Anecdotally:

We also ran a quick after-event survey (n=26):

Also, some random pictures:

Acknowledgements

We want to thank the following people for their amazing help:

  1. ^

    We might or might have not chosen to write South America instead of Latin America to conveniently exclude Mexico

  2. ^

    Especially considering that participating in-person meant showing up at 9:00  of a summer vacation day in an inconveniently-located venue.

  3. ^
  4. ^

    Adapted from this idea by Caroline Jeanmaire. 

  5. ^

     Taken from this idea by @Sabrina Zaki, Luke Ring and Aleks Baskakovs, that in turn comes from their (winning) proposal for Apart Research's language model hackathon.

  6. ^

     Inspired the ideas in @Geoffrey Irving and Amanda Askell's AI Safety Needs Social Scientists and Riccardo Volpato's Research ideas to study humans with AI Safety in Mind.

  7. ^

    People from CS Engineering take common courses in engineering, but are otherwise similar to CS.

  8. ^

    This was explicitly defined in contrast to Ethics of AI or similar fields.

  9. ^

    A Wilcoxon signed-rank test showed a statistically significant increase in interest in AI (W = 115.5, p < 0.001, r = 0.32) as well as AI Safety (W = 190.0, p < 0.001, r = 0.54). This was based on retrospective Likert-style questions on interest. It was hard to account for survey effects.