Feb 10, 2023
-
Feb 13, 2023
Online & In-Person
Scale Oversight for Machine Learning Hackathon




Join us for the fifth Alignment Jam where we get to spend 48 hours of intense research on how we can measure and monitor the safety of large-scale machine learning models. Work on safety benchmarks, models detecting faults in other models, self-monitoring systems , and so much else!
00:00:00:00
00:00:00:00
00:00:00:00
00:00:00:00
Join us for the fifth Alignment Jam where we get to spend 48 hours of intense research on how we can measure and monitor the safety of large-scale machine learning models. Work on safety benchmarks, models detecting faults in other models, self-monitoring systems , and so much else!
This event is ongoing.
This event has concluded.
24
Sign Ups
Overview
Resources
Overview

Hosted by Esben Kran, pseudobison, Zaki, fbarez, ruiqi-zhong · #alignmentjam
Join us for the fifth Alignment Jam where we get to spend 48 hours of intense research on how we can measure and monitor the safety of large-scale machine learning models. Work on safety benchmarks, models detecting faults in other models, self-monitoring systems , and so much else!
🏆$2,000 on the line

Measuring and monitoring safety
To make sure large machine learning models follow what we want them to do, we have to have people monitoring their safety. BUT, it is indeed very hard for just one person to monitor all the outputs of ChatGPT...
The objective of this hackathon is to research scalable solutions to this problem!
Can we create good benchmarks that run independently of human oversight?
Can we train AI models themselves to find faults in other models?
Can we create ways for one human to monitor a much larger amount of data?
Can we reduce the misgeneralization of the original model using some novel method?
These are all very interesting questions that we're excited to see your answers for during theses 48 hours
Reading group
Join the Discord above to be a part of the reading group where we read up on the research within scaling oversight! The current pieces are:
24
Sign Ups
Overview
Resources
Overview

Hosted by Esben Kran, pseudobison, Zaki, fbarez, ruiqi-zhong · #alignmentjam
Join us for the fifth Alignment Jam where we get to spend 48 hours of intense research on how we can measure and monitor the safety of large-scale machine learning models. Work on safety benchmarks, models detecting faults in other models, self-monitoring systems , and so much else!
🏆$2,000 on the line

Measuring and monitoring safety
To make sure large machine learning models follow what we want them to do, we have to have people monitoring their safety. BUT, it is indeed very hard for just one person to monitor all the outputs of ChatGPT...
The objective of this hackathon is to research scalable solutions to this problem!
Can we create good benchmarks that run independently of human oversight?
Can we train AI models themselves to find faults in other models?
Can we create ways for one human to monitor a much larger amount of data?
Can we reduce the misgeneralization of the original model using some novel method?
These are all very interesting questions that we're excited to see your answers for during theses 48 hours
Reading group
Join the Discord above to be a part of the reading group where we read up on the research within scaling oversight! The current pieces are:
24
Sign Ups
Overview
Resources
Overview

Hosted by Esben Kran, pseudobison, Zaki, fbarez, ruiqi-zhong · #alignmentjam
Join us for the fifth Alignment Jam where we get to spend 48 hours of intense research on how we can measure and monitor the safety of large-scale machine learning models. Work on safety benchmarks, models detecting faults in other models, self-monitoring systems , and so much else!
🏆$2,000 on the line

Measuring and monitoring safety
To make sure large machine learning models follow what we want them to do, we have to have people monitoring their safety. BUT, it is indeed very hard for just one person to monitor all the outputs of ChatGPT...
The objective of this hackathon is to research scalable solutions to this problem!
Can we create good benchmarks that run independently of human oversight?
Can we train AI models themselves to find faults in other models?
Can we create ways for one human to monitor a much larger amount of data?
Can we reduce the misgeneralization of the original model using some novel method?
These are all very interesting questions that we're excited to see your answers for during theses 48 hours
Reading group
Join the Discord above to be a part of the reading group where we read up on the research within scaling oversight! The current pieces are:
24
Sign Ups
Overview
Resources
Overview

Hosted by Esben Kran, pseudobison, Zaki, fbarez, ruiqi-zhong · #alignmentjam
Join us for the fifth Alignment Jam where we get to spend 48 hours of intense research on how we can measure and monitor the safety of large-scale machine learning models. Work on safety benchmarks, models detecting faults in other models, self-monitoring systems , and so much else!
🏆$2,000 on the line

Measuring and monitoring safety
To make sure large machine learning models follow what we want them to do, we have to have people monitoring their safety. BUT, it is indeed very hard for just one person to monitor all the outputs of ChatGPT...
The objective of this hackathon is to research scalable solutions to this problem!
Can we create good benchmarks that run independently of human oversight?
Can we train AI models themselves to find faults in other models?
Can we create ways for one human to monitor a much larger amount of data?
Can we reduce the misgeneralization of the original model using some novel method?
These are all very interesting questions that we're excited to see your answers for during theses 48 hours
Reading group
Join the Discord above to be a part of the reading group where we read up on the research within scaling oversight! The current pieces are:
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Our Other Sprints
May 30, 2025
-
Jun 1, 2025
Research
Apart x Martian Mechanistic Interpretability Hackathon
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up
Apr 25, 2025
-
Apr 27, 2025
Research
Economics of Transformative AI
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events