$250K in Prizes: SafeBench Competition Announcement

By Center for AI Safety @ 2024-04-03T22:07 (+47)

TLDR: CAIS is distributing $250,000 in prizes for benchmarks that empirically assess AI safety. This project is supported by Schmidt Sciences, submissions are open until February 25th, 2025. Winners will be announced April 25th, 2025.

To view additional info about the competition, including submission guidelines and FAQs, visit https://www.mlsafety.org/safebench 

If you are interested in receiving updates about SafeBench, feel free to sign up on the homepage using the link above.
 

About the Competition:

The Center for AI Safety is offering prizes for the best benchmarks across the following four categories:
 

Judges:

Timeline:

Mar 25, 2024: Competition Opens

Feb 25, 2025: Submission

Deadline Apr 25, 2025: Winners Announced

 

Competition Details:

Prizes: There will be three prizes worth $50,000 and five prizes worth $20,000. 

Eligibility: Benchmarks released prior to the competition launch are ineligible for prize consideration. Benchmarks released after competition launch are eligible. More details about prize eligibility can be found in our terms and conditions.

Evaluation criteria: Benchmarks will be assessed according to this evaluation criteria. In order to encourage progress in safety, without also encouraging general advances in capabilities, benchmarks must clearly delineate safety and capabilities.

Submission Format: If you have already written a paper on your benchmark, submit that (as long as it was published after the SafeBench launch date of March 25th, 2024). Otherwise, you may submit a thorough write-up of your benchmark, including source code. An example of such a write-up can be found in this document.

Dataset Policy: By default, we will require the code and dataset for all submissions to be publicly available on Github. However, if the submission deals with a dangerous capability, we will review whether to publicly release the dataset on a case-by-case basis.

If you are interested in receiving updates about SafeBench, feel free to sign up on the homepage: https://www.mlsafety.org/safebench