Funding Case: AI Safety Camp 11

By Remmelt, Linda Linsefors, Robert Kralisch @ 2024-12-23T08:39 (+42)

This is a linkpost to https://manifund.org/projects/11th-edition-of-ai-safety-camp

Project summary

AI Safety Camp has a seven-year track record of enabling participants to try their fit, find careers and start new orgs in AI Safety. We host up-and-coming researchers outside the Bay Area and London hubs.

If this fundraiser passes…

What are this project's goals? How will you achieve them?

By all accounts they are the gold standard for this type of thing. Everyone says they are great, I am generally a fan of the format, I buy that this can punch way above its weight or cost. If I was going to back [a talent funnel], I’d start here.

Zvi Mowshowitz (Nov 2024)

 

My current work (AI Standards Lab) was originally a AISC project. Without it, I'd guess I would be full-time employed in the field at least 1 year later, and the EU standards currently close to completion would be a lot weaker. High impact/high neglectedness opportunities are fairly well positioned to be kickstarted with volunteer effort in AISC, even if some projects will fail (hits based). After some initial results during AISC, they can be funded more easily.
Ariel Gil (Jan 2025)


AI Safety Camp is part incubator and part talent funnel:

The Incubator case is that AISC seeds epistemically diverse initiatives. Edition 10 supports new alignment directions, control limits research, neglected legal regulation, and 'slow down AI' advocacy. Funders who are uncertain on approaches to alignment – or believe we cannot align AGI on time – may prioritise funding this program.

The Maintaining Talent Funnels case is to give some money just to sustain the program. AISC is no longer the sole program training collaborators new to the field. There are now many programs, and our community’s bottlenecks have shifted to salary funding and org management. Still, new talent will be needed. For them, we can run a cost-efficient program. Sustaining this program retains optionality – institutions are waking up to AI risks and could greatly increase funding and positions there. If AISC still exists, it can help funnel people with a security mindset into those positions. But if by then organisers have left to new jobs, others would have to build AISC up from scratch. The cost of restarting is higher than it is to keep the program running.

As a funder, you may decide that AISC is worth saving as a cost-efficient talent funnel. Or you may decide that AISC is uniquely open to supporting unconventional approaches, and that something unexpectedly valuable may come out. 

Our program is both cost-efficient and scalable.

How will this funding be used?

Grant funding is tight. Without private donors, we cannot continue this program.

 

$15k: we won’t run a full program, but can facilitate 10 projects and preserve organising capabilities.

If we raise $15k, we won't run a full official edition. 

We can still commit to facilitating projects. Robert and Remmelt are already supporting projects in their respective fields of work. Robert has collaborated with other independent alignment researchers, as well as informally mentoring junior researchers doing conceptual and technical research on interpretable AI. Remmelt is kickstarting projects to slow down AI (eg. formalization work, MILD, Stop AI, inter-community calls, film by an award-winning director). 

We might each just support projects independently. Or we could (also) run an informal event where we only invite past alumni to collaborate on projects together. 

We can commit to this if we are freed from needing to transition to new jobs in 2025. Then we can resume full editions when grantmakers make more funds available. With a basic income of $18k each, we can commit to starting, mentoring, and/or coordinating 10 projects. 
 

$40k: we can organise the 11th edition, for 25 projects.

Combined with surplus funds from past camps (conservatively estimated at $21k), this covers salaries to Robert and Remmelt of $30.5k each. 

That is enough for us to organise the 11th edition. However, since we’d miss a third organiser, we’d only commit to hosting 25 projects. 
 

$70k: we can pay a third organiser, for 35 projects.

With funding, we are confident that we can onboard a new organiser to trial with us. They would assist Robert with evaluating technical safety proposals, and help with event ops. This gives us capacity to host 35 projects.
 

$300k: we can cover stipends for 40 projects.

Stipends act as a commitment device, and enable young researchers to focus on research without having to take on side-gigs. We only offer stipends to participants who indicate it would help their work. Our stipends are $1.5k per research lead and $1k per team member, plus admin fees of 9%.

We would pay out stipends in the following order:

The $230k extra safely covers stipends for edition 11. This amount may seem high, but it cost-efficiently supports 150+ people's work over three months. This in turn reduces the load on us organisers, allowing us to host 40 projects.
 

Who is on your team? 

Remmelt is coordinator of 'Stop/Pause AI'  projects:


Robert is coordinator of 'Conceptual and Technical AI Safety Research' projects:

Linda will take a break from organising, staying on as an advisor. We can hire a third organiser to take up her tasks.
 

What's your track record?

AI Safety Camp is primarily a learning-by-doing training program. People get to try a role and explore directions in AI safety, by collaborating on a concrete project.

Multiple alumni have told us that AI Safety Camp was how they got started in AI Safety.

Papers that came out of the camp include:

Projects started at AI Safety Camp went on to receive a total of $1.4 million in grants:
   AISC 1:  Bounded Rationality team    
     $30k from Paul
   AISC 3:  Modelling Cooperation
     $24k from CLT, $50k from SFF, $83k from SFF, $83k from SFF
   AISC 4:  Survey      
     $5k from LTTF
   AISC 5:  Pessimistic Agents      
     $3k from LTFF
   AISC 5:  Multi-Objective Alignment
     $20k from EV, $26k from LTFF
   AISC 6:  LMs as Tools for Alignment
     $10K from LTFF
   AISC 6:  Modularity
     $125k from LTFF
   AISC 7:  AGI Inherent Non-Safety
     $170k from SFF, $135k from SFF 
    AISC 8:  Policy Proposals for High-Risk AI     
     $10k from NL, $184k from SFF, $200k from OpenPhil, $200k from AISTOF 
    AISC 9:  Data Disclosure
      $10k from SFFsg
    AISC 9:  VAISU
      $10k from LTFF

Organizations launched out of camp conversations include:

Alumni went on to take positions at:

For statistics of previous editions, see here.
 

What are the most likely causes and outcomes if this project fails?

Not receiving minimum funding:

Projects are low priority:

Projects support capability work:

How much money have you raised in the last 12 months, and from where?


gergo @ 2025-01-09T16:21 (+3)

Someone mentioned to me that the problem they see with AISC is that its scope is way too broad. 

I think it's great that you can accommodate a lot of different projects, but I would guess this does make it harder for you to make the case for funding. 

Would you consider giving the option to funders to only fund certain types of projects? I could imagine many people wanting to fund technical research, but not advocacy, and vice versa. 

Remmelt @ 2025-01-18T06:03 (+3)

We ended up having a private exchange about it. 

Basically, organisers spend more than half of their time on general communications and logistics to support participants get to work. 

And earmarking stipends to particular areas of work seems rather burdensome administratively, though I wouldn’t be entirely against it if it means we can cover more people’s stipends.

Overall, I think we tended not to allow differentiated fundraising before because it can promote internal conflicts, rather than having people come together to make the camp great.