Aug 23, 2024
-
Aug 26, 2024
AI capabilities and risks demo-jam
As frontier AI systems become rapidly more powerful and general, and as their risks become more pressing, it’s increasingly important for key decision-makers and the public to understand their capabilities. Let’s bring these insights out of dusty Arxiv papers and build visceral, immediately engaging demos!
This event is ongoing.
This event has concluded.
As frontier AI systems become rapidly more powerful and general, and as their risks become more pressing, it’s increasingly important for key decision-makers and the public to understand their capabilities. Let’s bring these insights out of dusty Arxiv papers and build visceral, immediately engaging demos!
Well-crafted interactive demos can be incredibly powerful in conveying AI capabilities and risks. That's why we're inviting you – developers, designers, and AI enthusiasts – to team up and create innovative demos that make people feel, rather than just think, about the rate of AI progress.
Why interactive demos can have a drastic impact
Interactive demonstrations allow something that just reading about AI safety can't do: make the reader live through what potential scenarios could look like, and engage them with complex concepts firsthand.
Importantly, what people understand and think has an impact on political decisions: “The Impact of Public Opinion on Public Policy: A Review and an Agenda” finds that public opinion plays a key-role in public policy, and the more so the more salient this opinion is, while “Does Public Opinion Affect Political Speech?” finds that politicians adjust their speech and position to reflect the preference of the public. This is why to impact AI policy we need both communication with politicians but also to communicate ideas with people at large.
We are focusing on interactive demos as an underappreciated way to communicate on an issue, as interactivity allows one to truly live through the concept presented and get an intuitive feel for it.
For this hackathon, we’re excited about the following possibilities:
Present scenarios that allow users to emotionally connect with potential AI risks
Illustrate key AI safety concepts and trends in an engaging, hands-on manner
Showcase current AI capabilities and their implications in a tangible way
Some examples of demonstrations we would be excited to see as a result of this hackathon:
How can AI disrupt elections? (AI Digest): Have a call with an AI that wants to prevent you from voting
AI-Powered Cybersecurity Threats (CivAI): Experience what automated social-engineering assisted by AI would look like
FoxVox (Palisade Research): See how AI can use subtle rewording to have an impact on your political views of recent events
Take a look at our resources to see many more!
What to expect during a demo-jam hackathon?
The hackathon is a weekend-long event where you participate in teams (1-5) to create interactive applications that demonstrate AI risks, AI safety concepts, or current capabilities.
You will have the opportunity to:
Collaborate with like-minded individuals passionate about AI safety
Receive continuous feedback from other participants and mentors
Review and learn from other teams' projects
Contribute to raising awareness about crucial AI safety issues
Prizes, evaluation, and submission
During this hackathon, you will develop and submit an interactive application, as well as review other participants’ entries.
You’ll submit:
source code of the application, along with instructions to deploy it locally.
You can use any tools or language as long as the deployment information are simple to follow
Be sure to remove your API keys before submission.
We recommend tools such as docker-compose for participants to easily and safely judge your submission.
We also recommend using either a public-weight models, or AI tools that have enough of a free trial for participants to test your submission
a brief report summarizing your project following this template
a 2-minute video demonstration of how your interactive demo works
Optionally, an endpoint for participants to test your application directly (for example through Vercel).
The submission will be reviewed under the following criteria:
Viscerality: Does the demo give you a felt sense of AI capabilities? Is it fun, engaging, immersive or understandable? Does it make you feel the real-world implications?
Importance: How important are the insights delivered by this demo for understanding how AI will affect humanity in the coming years?
Accuracy: If the demo shows capabilities or risks, does it present them accurately, giving takeaways that generalize well? If the demonstration explains a concept, is it accurate?
Ease of testing: How easy is it to try out the demo?
Safety: Is the demo safe enough to publish publicly without causing harm?
Overall: How good is the demo overall?
The judging panel's scores will count for half of the final score, with the other half coming from peer reviews. You will be assigned as a peer reviewer of some other hackathon submissions - you must review these to be eligible to win.
Top teams will win a share of our $2,000 prize pool:
🥇 1st place: $1,000
🥈 2nd place: $600
🥉 3rd place: $300
🎖️ 4th place: $100
Additionally, high quality submissions that are a good fit for AI Digest might be invited to collaborate to polish up and publish their demo there, to reach a wider audience of policymakers and the public.
Why should I join?
There’s loads of reasons to join! Here are just a few:
Participate in a game jam-style event focused on an AI safety-adjacent topic
Get to know new people interested in AI-safety
Win up to $1,000
Receive a certificate of participation
Contribute to important public outreach on AI safety
And many many more… Come along!
Do I need experience in AI safety to join?
Not at all! This can be an occasion for you to learn more about AI safety and public outreach. We provide code templates and ideas to kickstart your projects, mentors to give feedback on your project, and a great community of interested developers to give reviews and feedback on your project.
What if my project seems too risky to share?
Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.
For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy
I have more questions to ask
Head over to the Frequently Asked Questions on the submission tab, and feel free to ping us ( @mentor ) on Discord!
Resources
Hackbook: Basic code template to jump-start your submission
AI risk demonstrations:
How can AI disrupt elections? (AI Digest): Talk to an AI robocaller in this interactive demo.
TerifAI (Aman Ibrahim): An AI that steals your voice and impersonates you after interacting with it.
Deepfakes Sandbox (CivAI): Automated deepfake generator to demonstrate how easy creating a deepfake of someone is currently.
AI-Powered Cybersecurity Threats (CivAI): Demonstration of what automated phishing and impersonation will look like.
FoxVox (Palisade Research): Automated subtle rewording for political bias.
AI capabilities and trends demonstrations:
How fast is AI improving? (AI Digest): Interactive explainer about past AI trends.
GPT-4 capability forecasting calibration game (Carlini): Calibrate to see what GPT-4 can and can not do.
Claude 3 vs Gemini Ultra vs GPT-4 (AI Digest): A side-by-side comparison
Timeline of AI forecasts (AI Digest): What to expect in AI, according to forecasters.
Human or Not: Turing test game to see how good AIs are at passing off for humans.
What makes a good interactive demos?
Nicky Case’s tutorial for how to do good explorable explanations
Evolution of trust (Nicky Case): A non AI demonstration teaching principles of game theory.
Explorable explanations: Collections of interactive concept explanations.
Making an interactive application
Examples of AI tools you can use in your demo
Langchain: Library to call LLMs and has a universal interface, allowing you to easily switch to other models.
LLama3.1: A performant open-weight LLM model. We recommend using open-weight models for participant to more easily judge your demo.
Claude Sonnet 3.5: A private LLM model that is cheap and performant
GPT-4o: Private and performant LLM model
Retell.ai: A service for automated call platforms
Openrouter: Universal interface for LLM through API call
The schedule runs from 4 PM UTC Friday to 3 AM Monday. We start with an introductory talk and end the event during the following week with an awards ceremony. Join the public ICal here.
You will also find Explorer events, such as collaborative brainstorming and team match-making before the hackathon begins on Discord and in the calendar.

Entries
Our Other Sprints
Apr 25, 2025
-
Apr 27, 2025
Economics of Transformative AI: Research Sprint
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Apr 25, 2025
-
Apr 26, 2025
Berkeley AI Policy Hackathon
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible