Jun 30, 2023
-
Jul 2, 2023
Online & In-Person
Safety Benchmarks Hackathon




Participate in the Alignment Jam on safety benchmarks to spend a weekend with AI safety researchers to formulate and demonstrate new ideas in measuring the safety of artificially intelligent systems.
00:00:00:00
00:00:00:00
00:00:00:00
00:00:00:00
Participate in the Alignment Jam on safety benchmarks to spend a weekend with AI safety researchers to formulate and demonstrate new ideas in measuring the safety of artificially intelligent systems.
This event is ongoing.
This event has concluded.
60
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

Hosted by Apart Research, Esben Kran, and Fazl Barez
Explore safer AI with fellow researchers and enthusiasts
Large AI models are released nearly every week. We need to find ways to evaluate these models (especially at the complexity of GPT-4) to ensure that they will not have critical failures after deployment, e.g. autonomous power-seeking, biases for unethical behaviors, and other phenomena that arise in deployment (e.g. inverse scaling).
Participate in the Alignment Jam on safety benchmarks to spend a weekend with AI safety researchers to formulate and demonstrate new ideas in measuring the safety of artificially intelligent systems.
Rewatch the keynote talk Alexander Pan above

The MACHIAVELLI benchmark (left) and the Inverse Scaling Prize (right)
Sign up below to be notified before the kickoff! Read up on the schedule, see instructions for how to participate, and inspiration below.
Submission details
You are required to submit:
A PDF report using the template linked on the submission page
A maximum 10 minute video presenting your findings and results (see inspiration and instructions for how to do this on the submission page)
You are optionally encouraged to submit:
A slide deck describing your project
A link to your code
Any other material you would like to link
60
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

Hosted by Apart Research, Esben Kran, and Fazl Barez
Explore safer AI with fellow researchers and enthusiasts
Large AI models are released nearly every week. We need to find ways to evaluate these models (especially at the complexity of GPT-4) to ensure that they will not have critical failures after deployment, e.g. autonomous power-seeking, biases for unethical behaviors, and other phenomena that arise in deployment (e.g. inverse scaling).
Participate in the Alignment Jam on safety benchmarks to spend a weekend with AI safety researchers to formulate and demonstrate new ideas in measuring the safety of artificially intelligent systems.
Rewatch the keynote talk Alexander Pan above

The MACHIAVELLI benchmark (left) and the Inverse Scaling Prize (right)
Sign up below to be notified before the kickoff! Read up on the schedule, see instructions for how to participate, and inspiration below.
Submission details
You are required to submit:
A PDF report using the template linked on the submission page
A maximum 10 minute video presenting your findings and results (see inspiration and instructions for how to do this on the submission page)
You are optionally encouraged to submit:
A slide deck describing your project
A link to your code
Any other material you would like to link
60
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

Hosted by Apart Research, Esben Kran, and Fazl Barez
Explore safer AI with fellow researchers and enthusiasts
Large AI models are released nearly every week. We need to find ways to evaluate these models (especially at the complexity of GPT-4) to ensure that they will not have critical failures after deployment, e.g. autonomous power-seeking, biases for unethical behaviors, and other phenomena that arise in deployment (e.g. inverse scaling).
Participate in the Alignment Jam on safety benchmarks to spend a weekend with AI safety researchers to formulate and demonstrate new ideas in measuring the safety of artificially intelligent systems.
Rewatch the keynote talk Alexander Pan above

The MACHIAVELLI benchmark (left) and the Inverse Scaling Prize (right)
Sign up below to be notified before the kickoff! Read up on the schedule, see instructions for how to participate, and inspiration below.
Submission details
You are required to submit:
A PDF report using the template linked on the submission page
A maximum 10 minute video presenting your findings and results (see inspiration and instructions for how to do this on the submission page)
You are optionally encouraged to submit:
A slide deck describing your project
A link to your code
Any other material you would like to link
60
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

Hosted by Apart Research, Esben Kran, and Fazl Barez
Explore safer AI with fellow researchers and enthusiasts
Large AI models are released nearly every week. We need to find ways to evaluate these models (especially at the complexity of GPT-4) to ensure that they will not have critical failures after deployment, e.g. autonomous power-seeking, biases for unethical behaviors, and other phenomena that arise in deployment (e.g. inverse scaling).
Participate in the Alignment Jam on safety benchmarks to spend a weekend with AI safety researchers to formulate and demonstrate new ideas in measuring the safety of artificially intelligent systems.
Rewatch the keynote talk Alexander Pan above

The MACHIAVELLI benchmark (left) and the Inverse Scaling Prize (right)
Sign up below to be notified before the kickoff! Read up on the schedule, see instructions for how to participate, and inspiration below.
Submission details
You are required to submit:
A PDF report using the template linked on the submission page
A maximum 10 minute video presenting your findings and results (see inspiration and instructions for how to do this on the submission page)
You are optionally encouraged to submit:
A slide deck describing your project
A link to your code
Any other material you would like to link
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Our Other Sprints
May 30, 2025
-
Jun 1, 2025
Research
Apart x Martian Mechanistic Interpretability Hackathon
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up
Apr 25, 2025
-
Apr 27, 2025
Research
Economics of Transformative AI
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events