Aug 23, 2024
-
Aug 26, 2024
Online & In-Person
AI capabilities and risks demo-jam




As frontier AI systems become rapidly more powerful and general, and as their risks become more pressing, it’s increasingly important for key decision-makers and the public to understand their capabilities. Let’s bring these insights out of dusty Arxiv papers and build visceral, immediately engaging demos!
00:00:00:00
00:00:00:00
00:00:00:00
00:00:00:00
This event is ongoing.
This event has concluded.
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

As frontier AI systems become rapidly more powerful and general, and as their risks become more pressing, it’s increasingly important for key decision-makers and the public to understand their capabilities. Let’s bring these insights out of dusty Arxiv papers and build visceral, immediately engaging demos!
Well-crafted interactive demos can be incredibly powerful in conveying AI capabilities and risks. That's why we're inviting you – developers, designers, and AI enthusiasts – to team up and create innovative demos that make people feel, rather than just think, about the rate of AI progress.
Why interactive demos can have a drastic impact
Interactive demonstrations allow something that just reading about AI safety can't do: make the reader live through what potential scenarios could look like, and engage them with complex concepts firsthand.
Importantly, what people understand and think has an impact on political decisions: “The Impact of Public Opinion on Public Policy: A Review and an Agenda” finds that public opinion plays a key-role in public policy, and the more so the more salient this opinion is, while “Does Public Opinion Affect Political Speech?” finds that politicians adjust their speech and position to reflect the preference of the public. This is why to impact AI policy we need both communication with politicians but also to communicate ideas with people at large.
We are focusing on interactive demos as an underappreciated way to communicate on an issue, as interactivity allows one to truly live through the concept presented and get an intuitive feel for it.
For this hackathon, we’re excited about the following possibilities:
Present scenarios that allow users to emotionally connect with potential AI risks
Illustrate key AI safety concepts and trends in an engaging, hands-on manner
Showcase current AI capabilities and their implications in a tangible way
Some examples of demonstrations we would be excited to see as a result of this hackathon:
How can AI disrupt elections? (AI Digest): Have a call with an AI that wants to prevent you from voting
AI-Powered Cybersecurity Threats (CivAI): Experience what automated social-engineering assisted by AI would look like
FoxVox (Palisade Research): See how AI can use subtle rewording to have an impact on your political views of recent events
Take a look at our resources to see many more!
What to expect during a demo-jam hackathon?
The hackathon is a weekend-long event where you participate in teams (1-5) to create interactive applications that demonstrate AI risks, AI safety concepts, or current capabilities.
You will have the opportunity to:
Collaborate with like-minded individuals passionate about AI safety
Receive continuous feedback from other participants and mentors
Review and learn from other teams' projects
Contribute to raising awareness about crucial AI safety issues
Prizes, evaluation, and submission
During this hackathon, you will develop and submit an interactive application, as well as review other participants’ entries.
You’ll submit:
source code of the application, along with instructions to deploy it locally.
You can use any tools or language as long as the deployment information are simple to follow
Be sure to remove your API keys before submission.
We recommend tools such as docker-compose for participants to easily and safely judge your submission.
We also recommend using either a public-weight models, or AI tools that have enough of a free trial for participants to test your submission
a brief report summarizing your project following this template
a 2-minute video demonstration of how your interactive demo works
Optionally, an endpoint for participants to test your application directly (for example through Vercel).
The submission will be reviewed under the following criteria:
Viscerality: Does the demo give you a felt sense of AI capabilities? Is it fun, engaging, immersive or understandable? Does it make you feel the real-world implications?
Importance: How important are the insights delivered by this demo for understanding how AI will affect humanity in the coming years?
Accuracy: If the demo shows capabilities or risks, does it present them accurately, giving takeaways that generalize well? If the demonstration explains a concept, is it accurate?
Ease of testing: How easy is it to try out the demo?
Safety: Is the demo safe enough to publish publicly without causing harm?
Overall: How good is the demo overall?
The judging panel's scores will count for half of the final score, with the other half coming from peer reviews. You will be assigned as a peer reviewer of some other hackathon submissions - you must review these to be eligible to win.
Top teams will win a share of our $2,000 prize pool:
🥇 1st place: $1,000
🥈 2nd place: $600
🥉 3rd place: $300
🎖️ 4th place: $100
Additionally, high quality submissions that are a good fit for AI Digest might be invited to collaborate to polish up and publish their demo there, to reach a wider audience of policymakers and the public.
Why should I join?
There’s loads of reasons to join! Here are just a few:
Participate in a game jam-style event focused on an AI safety-adjacent topic
Get to know new people interested in AI-safety
Win up to $1,000
Receive a certificate of participation
Contribute to important public outreach on AI safety
And many many more… Come along!
Do I need experience in AI safety to join?
Not at all! This can be an occasion for you to learn more about AI safety and public outreach. We provide code templates and ideas to kickstart your projects, mentors to give feedback on your project, and a great community of interested developers to give reviews and feedback on your project.
What if my project seems too risky to share?
Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.
For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy
I have more questions to ask
Head over to the Frequently Asked Questions on the submission tab, and feel free to ping us ( @mentor ) on Discord!
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

As frontier AI systems become rapidly more powerful and general, and as their risks become more pressing, it’s increasingly important for key decision-makers and the public to understand their capabilities. Let’s bring these insights out of dusty Arxiv papers and build visceral, immediately engaging demos!
Well-crafted interactive demos can be incredibly powerful in conveying AI capabilities and risks. That's why we're inviting you – developers, designers, and AI enthusiasts – to team up and create innovative demos that make people feel, rather than just think, about the rate of AI progress.
Why interactive demos can have a drastic impact
Interactive demonstrations allow something that just reading about AI safety can't do: make the reader live through what potential scenarios could look like, and engage them with complex concepts firsthand.
Importantly, what people understand and think has an impact on political decisions: “The Impact of Public Opinion on Public Policy: A Review and an Agenda” finds that public opinion plays a key-role in public policy, and the more so the more salient this opinion is, while “Does Public Opinion Affect Political Speech?” finds that politicians adjust their speech and position to reflect the preference of the public. This is why to impact AI policy we need both communication with politicians but also to communicate ideas with people at large.
We are focusing on interactive demos as an underappreciated way to communicate on an issue, as interactivity allows one to truly live through the concept presented and get an intuitive feel for it.
For this hackathon, we’re excited about the following possibilities:
Present scenarios that allow users to emotionally connect with potential AI risks
Illustrate key AI safety concepts and trends in an engaging, hands-on manner
Showcase current AI capabilities and their implications in a tangible way
Some examples of demonstrations we would be excited to see as a result of this hackathon:
How can AI disrupt elections? (AI Digest): Have a call with an AI that wants to prevent you from voting
AI-Powered Cybersecurity Threats (CivAI): Experience what automated social-engineering assisted by AI would look like
FoxVox (Palisade Research): See how AI can use subtle rewording to have an impact on your political views of recent events
Take a look at our resources to see many more!
What to expect during a demo-jam hackathon?
The hackathon is a weekend-long event where you participate in teams (1-5) to create interactive applications that demonstrate AI risks, AI safety concepts, or current capabilities.
You will have the opportunity to:
Collaborate with like-minded individuals passionate about AI safety
Receive continuous feedback from other participants and mentors
Review and learn from other teams' projects
Contribute to raising awareness about crucial AI safety issues
Prizes, evaluation, and submission
During this hackathon, you will develop and submit an interactive application, as well as review other participants’ entries.
You’ll submit:
source code of the application, along with instructions to deploy it locally.
You can use any tools or language as long as the deployment information are simple to follow
Be sure to remove your API keys before submission.
We recommend tools such as docker-compose for participants to easily and safely judge your submission.
We also recommend using either a public-weight models, or AI tools that have enough of a free trial for participants to test your submission
a brief report summarizing your project following this template
a 2-minute video demonstration of how your interactive demo works
Optionally, an endpoint for participants to test your application directly (for example through Vercel).
The submission will be reviewed under the following criteria:
Viscerality: Does the demo give you a felt sense of AI capabilities? Is it fun, engaging, immersive or understandable? Does it make you feel the real-world implications?
Importance: How important are the insights delivered by this demo for understanding how AI will affect humanity in the coming years?
Accuracy: If the demo shows capabilities or risks, does it present them accurately, giving takeaways that generalize well? If the demonstration explains a concept, is it accurate?
Ease of testing: How easy is it to try out the demo?
Safety: Is the demo safe enough to publish publicly without causing harm?
Overall: How good is the demo overall?
The judging panel's scores will count for half of the final score, with the other half coming from peer reviews. You will be assigned as a peer reviewer of some other hackathon submissions - you must review these to be eligible to win.
Top teams will win a share of our $2,000 prize pool:
🥇 1st place: $1,000
🥈 2nd place: $600
🥉 3rd place: $300
🎖️ 4th place: $100
Additionally, high quality submissions that are a good fit for AI Digest might be invited to collaborate to polish up and publish their demo there, to reach a wider audience of policymakers and the public.
Why should I join?
There’s loads of reasons to join! Here are just a few:
Participate in a game jam-style event focused on an AI safety-adjacent topic
Get to know new people interested in AI-safety
Win up to $1,000
Receive a certificate of participation
Contribute to important public outreach on AI safety
And many many more… Come along!
Do I need experience in AI safety to join?
Not at all! This can be an occasion for you to learn more about AI safety and public outreach. We provide code templates and ideas to kickstart your projects, mentors to give feedback on your project, and a great community of interested developers to give reviews and feedback on your project.
What if my project seems too risky to share?
Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.
For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy
I have more questions to ask
Head over to the Frequently Asked Questions on the submission tab, and feel free to ping us ( @mentor ) on Discord!
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

As frontier AI systems become rapidly more powerful and general, and as their risks become more pressing, it’s increasingly important for key decision-makers and the public to understand their capabilities. Let’s bring these insights out of dusty Arxiv papers and build visceral, immediately engaging demos!
Well-crafted interactive demos can be incredibly powerful in conveying AI capabilities and risks. That's why we're inviting you – developers, designers, and AI enthusiasts – to team up and create innovative demos that make people feel, rather than just think, about the rate of AI progress.
Why interactive demos can have a drastic impact
Interactive demonstrations allow something that just reading about AI safety can't do: make the reader live through what potential scenarios could look like, and engage them with complex concepts firsthand.
Importantly, what people understand and think has an impact on political decisions: “The Impact of Public Opinion on Public Policy: A Review and an Agenda” finds that public opinion plays a key-role in public policy, and the more so the more salient this opinion is, while “Does Public Opinion Affect Political Speech?” finds that politicians adjust their speech and position to reflect the preference of the public. This is why to impact AI policy we need both communication with politicians but also to communicate ideas with people at large.
We are focusing on interactive demos as an underappreciated way to communicate on an issue, as interactivity allows one to truly live through the concept presented and get an intuitive feel for it.
For this hackathon, we’re excited about the following possibilities:
Present scenarios that allow users to emotionally connect with potential AI risks
Illustrate key AI safety concepts and trends in an engaging, hands-on manner
Showcase current AI capabilities and their implications in a tangible way
Some examples of demonstrations we would be excited to see as a result of this hackathon:
How can AI disrupt elections? (AI Digest): Have a call with an AI that wants to prevent you from voting
AI-Powered Cybersecurity Threats (CivAI): Experience what automated social-engineering assisted by AI would look like
FoxVox (Palisade Research): See how AI can use subtle rewording to have an impact on your political views of recent events
Take a look at our resources to see many more!
What to expect during a demo-jam hackathon?
The hackathon is a weekend-long event where you participate in teams (1-5) to create interactive applications that demonstrate AI risks, AI safety concepts, or current capabilities.
You will have the opportunity to:
Collaborate with like-minded individuals passionate about AI safety
Receive continuous feedback from other participants and mentors
Review and learn from other teams' projects
Contribute to raising awareness about crucial AI safety issues
Prizes, evaluation, and submission
During this hackathon, you will develop and submit an interactive application, as well as review other participants’ entries.
You’ll submit:
source code of the application, along with instructions to deploy it locally.
You can use any tools or language as long as the deployment information are simple to follow
Be sure to remove your API keys before submission.
We recommend tools such as docker-compose for participants to easily and safely judge your submission.
We also recommend using either a public-weight models, or AI tools that have enough of a free trial for participants to test your submission
a brief report summarizing your project following this template
a 2-minute video demonstration of how your interactive demo works
Optionally, an endpoint for participants to test your application directly (for example through Vercel).
The submission will be reviewed under the following criteria:
Viscerality: Does the demo give you a felt sense of AI capabilities? Is it fun, engaging, immersive or understandable? Does it make you feel the real-world implications?
Importance: How important are the insights delivered by this demo for understanding how AI will affect humanity in the coming years?
Accuracy: If the demo shows capabilities or risks, does it present them accurately, giving takeaways that generalize well? If the demonstration explains a concept, is it accurate?
Ease of testing: How easy is it to try out the demo?
Safety: Is the demo safe enough to publish publicly without causing harm?
Overall: How good is the demo overall?
The judging panel's scores will count for half of the final score, with the other half coming from peer reviews. You will be assigned as a peer reviewer of some other hackathon submissions - you must review these to be eligible to win.
Top teams will win a share of our $2,000 prize pool:
🥇 1st place: $1,000
🥈 2nd place: $600
🥉 3rd place: $300
🎖️ 4th place: $100
Additionally, high quality submissions that are a good fit for AI Digest might be invited to collaborate to polish up and publish their demo there, to reach a wider audience of policymakers and the public.
Why should I join?
There’s loads of reasons to join! Here are just a few:
Participate in a game jam-style event focused on an AI safety-adjacent topic
Get to know new people interested in AI-safety
Win up to $1,000
Receive a certificate of participation
Contribute to important public outreach on AI safety
And many many more… Come along!
Do I need experience in AI safety to join?
Not at all! This can be an occasion for you to learn more about AI safety and public outreach. We provide code templates and ideas to kickstart your projects, mentors to give feedback on your project, and a great community of interested developers to give reviews and feedback on your project.
What if my project seems too risky to share?
Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.
For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy
I have more questions to ask
Head over to the Frequently Asked Questions on the submission tab, and feel free to ping us ( @mentor ) on Discord!
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

As frontier AI systems become rapidly more powerful and general, and as their risks become more pressing, it’s increasingly important for key decision-makers and the public to understand their capabilities. Let’s bring these insights out of dusty Arxiv papers and build visceral, immediately engaging demos!
Well-crafted interactive demos can be incredibly powerful in conveying AI capabilities and risks. That's why we're inviting you – developers, designers, and AI enthusiasts – to team up and create innovative demos that make people feel, rather than just think, about the rate of AI progress.
Why interactive demos can have a drastic impact
Interactive demonstrations allow something that just reading about AI safety can't do: make the reader live through what potential scenarios could look like, and engage them with complex concepts firsthand.
Importantly, what people understand and think has an impact on political decisions: “The Impact of Public Opinion on Public Policy: A Review and an Agenda” finds that public opinion plays a key-role in public policy, and the more so the more salient this opinion is, while “Does Public Opinion Affect Political Speech?” finds that politicians adjust their speech and position to reflect the preference of the public. This is why to impact AI policy we need both communication with politicians but also to communicate ideas with people at large.
We are focusing on interactive demos as an underappreciated way to communicate on an issue, as interactivity allows one to truly live through the concept presented and get an intuitive feel for it.
For this hackathon, we’re excited about the following possibilities:
Present scenarios that allow users to emotionally connect with potential AI risks
Illustrate key AI safety concepts and trends in an engaging, hands-on manner
Showcase current AI capabilities and their implications in a tangible way
Some examples of demonstrations we would be excited to see as a result of this hackathon:
How can AI disrupt elections? (AI Digest): Have a call with an AI that wants to prevent you from voting
AI-Powered Cybersecurity Threats (CivAI): Experience what automated social-engineering assisted by AI would look like
FoxVox (Palisade Research): See how AI can use subtle rewording to have an impact on your political views of recent events
Take a look at our resources to see many more!
What to expect during a demo-jam hackathon?
The hackathon is a weekend-long event where you participate in teams (1-5) to create interactive applications that demonstrate AI risks, AI safety concepts, or current capabilities.
You will have the opportunity to:
Collaborate with like-minded individuals passionate about AI safety
Receive continuous feedback from other participants and mentors
Review and learn from other teams' projects
Contribute to raising awareness about crucial AI safety issues
Prizes, evaluation, and submission
During this hackathon, you will develop and submit an interactive application, as well as review other participants’ entries.
You’ll submit:
source code of the application, along with instructions to deploy it locally.
You can use any tools or language as long as the deployment information are simple to follow
Be sure to remove your API keys before submission.
We recommend tools such as docker-compose for participants to easily and safely judge your submission.
We also recommend using either a public-weight models, or AI tools that have enough of a free trial for participants to test your submission
a brief report summarizing your project following this template
a 2-minute video demonstration of how your interactive demo works
Optionally, an endpoint for participants to test your application directly (for example through Vercel).
The submission will be reviewed under the following criteria:
Viscerality: Does the demo give you a felt sense of AI capabilities? Is it fun, engaging, immersive or understandable? Does it make you feel the real-world implications?
Importance: How important are the insights delivered by this demo for understanding how AI will affect humanity in the coming years?
Accuracy: If the demo shows capabilities or risks, does it present them accurately, giving takeaways that generalize well? If the demonstration explains a concept, is it accurate?
Ease of testing: How easy is it to try out the demo?
Safety: Is the demo safe enough to publish publicly without causing harm?
Overall: How good is the demo overall?
The judging panel's scores will count for half of the final score, with the other half coming from peer reviews. You will be assigned as a peer reviewer of some other hackathon submissions - you must review these to be eligible to win.
Top teams will win a share of our $2,000 prize pool:
🥇 1st place: $1,000
🥈 2nd place: $600
🥉 3rd place: $300
🎖️ 4th place: $100
Additionally, high quality submissions that are a good fit for AI Digest might be invited to collaborate to polish up and publish their demo there, to reach a wider audience of policymakers and the public.
Why should I join?
There’s loads of reasons to join! Here are just a few:
Participate in a game jam-style event focused on an AI safety-adjacent topic
Get to know new people interested in AI-safety
Win up to $1,000
Receive a certificate of participation
Contribute to important public outreach on AI safety
And many many more… Come along!
Do I need experience in AI safety to join?
Not at all! This can be an occasion for you to learn more about AI safety and public outreach. We provide code templates and ideas to kickstart your projects, mentors to give feedback on your project, and a great community of interested developers to give reviews and feedback on your project.
What if my project seems too risky to share?
Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.
For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy
I have more questions to ask
Head over to the Frequently Asked Questions on the submission tab, and feel free to ping us ( @mentor ) on Discord!
Speakers & Collaborators
Lucas Hansen
Keynote Speaker & Judge
Lucas brings 10 years of technology experience to CivAI. Previously, he founded Qualia, a real estate software company which serves millions of homeowners each year.
Épiphanie Gédéon
Organizer, Mentor & Judge
Épiphanie is a member of Centre pour la Sécurité de l'IA and currently an Apart Research Fellow
Adam Binksmith
Organizer, Mentor & Judge
Adam runs Sage, an organization helping impactful communities figure out what's true.
Archana Vaidheeswaran
Organizer
Archana is responsible for organizing the Apart Sprints, research hackathons to solve the most important questions in AI safety.
Misha Yagudin
Mentor & Judge
Misha runs Arb Research and Samotsvety Forecasting. Misha currently advises at Sage
Charbel-Raphaël Ségerie
Judge
Charbel-Raphaël is the executive director at Centre pour la Sécurité de l'IA and has led ML4Good, a bootcamp focused on AI safety.
Natalia Pérez-Campanero Antolín
Judge
A research manager at Apart, Natalia has a PhD in Interdisciplinary Biosciences from Oxford and has run the Royal Society's Entrepreneur-in-Residence program.
Jason Schreiber
Organizer and Judge
Jason is co-director of Apart Research and leads Apart Lab, our remote-first AI safety research fellowship.
Speakers & Collaborators

Lucas Hansen
Keynote Speaker & Judge
Lucas brings 10 years of technology experience to CivAI. Previously, he founded Qualia, a real estate software company which serves millions of homeowners each year.

Épiphanie Gédéon
Organizer, Mentor & Judge
Épiphanie is a member of Centre pour la Sécurité de l'IA and currently an Apart Research Fellow

Adam Binksmith
Organizer, Mentor & Judge
Adam runs Sage, an organization helping impactful communities figure out what's true.

Archana Vaidheeswaran
Organizer
Archana is responsible for organizing the Apart Sprints, research hackathons to solve the most important questions in AI safety.

Misha Yagudin
Mentor & Judge
Misha runs Arb Research and Samotsvety Forecasting. Misha currently advises at Sage

Charbel-Raphaël Ségerie
Judge
Charbel-Raphaël is the executive director at Centre pour la Sécurité de l'IA and has led ML4Good, a bootcamp focused on AI safety.

Natalia Pérez-Campanero Antolín
Judge
A research manager at Apart, Natalia has a PhD in Interdisciplinary Biosciences from Oxford and has run the Royal Society's Entrepreneur-in-Residence program.
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Our Other Sprints
Apr 25, 2025
-
Apr 27, 2025
Research
Economics of Transformative AI: Research Sprint
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up
Apr 25, 2025
-
Apr 26, 2025
Research
Berkeley AI Policy Hackathon
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events