Aug 23, 2024

-

Aug 26, 2024

Online & In-Person

AI capabilities and risks demo-jam

As frontier AI systems become rapidly more powerful and general, and as their risks become more pressing, it’s increasingly important for key decision-makers and the public to understand their capabilities. Let’s bring these insights out of dusty Arxiv papers and build visceral, immediately engaging demos!

00:00:00:00

00:00:00:00

00:00:00:00

00:00:00:00

This event is ongoing.

This event has concluded.

Overview

Overview

Arrow
Arrow
Arrow

As frontier AI systems become rapidly more powerful and general, and as their risks become more pressing, it’s increasingly important for key decision-makers and the public to understand their capabilities. Let’s bring these insights out of dusty Arxiv papers and build visceral, immediately engaging demos!

Well-crafted interactive demos can be incredibly powerful in conveying AI capabilities and risks. That's why we're inviting you – developers, designers, and AI enthusiasts – to team up and create innovative demos that make people feel, rather than just think, about the rate of AI progress.

Why interactive demos can have a drastic impact

Interactive demonstrations allow something that just reading about AI safety can't do: make the reader live through what potential scenarios could look like, and engage them with complex concepts firsthand.

Importantly, what people understand and think has an impact on political decisions: “The Impact of Public Opinion on Public Policy: A Review and an Agenda” finds that public opinion plays a key-role in public policy, and the more so the more salient this opinion is, while “Does Public Opinion Affect Political Speech?” finds that politicians adjust their speech and position to reflect the preference of the public. This is why to impact AI policy we need both communication with politicians but also to communicate ideas with people at large.

We are focusing on interactive demos as an underappreciated way to communicate on an issue, as interactivity allows one to truly live through the concept presented and get an intuitive feel for it.

For this hackathon, we’re excited about the following possibilities:

  • Present scenarios that allow users to emotionally connect with potential AI risks

  • Illustrate key AI safety concepts and trends in an engaging, hands-on manner

  • Showcase current AI capabilities and their implications in a tangible way

Some examples of demonstrations we would be excited to see as a result of this hackathon:

Take a look at our resources to see many more!

What to expect during a demo-jam hackathon?

The hackathon is a weekend-long event where you participate in teams (1-5) to create interactive applications that demonstrate AI risks, AI safety concepts, or current capabilities.

You will have the opportunity to:

  • Collaborate with like-minded individuals passionate about AI safety

  • Receive continuous feedback from other participants and mentors

  • Review and learn from other teams' projects

  • Contribute to raising awareness about crucial AI safety issues

Prizes, evaluation, and submission

During this hackathon, you will develop and submit an interactive application, as well as review other participants’ entries.

You’ll submit:

  • source code of the application, along with instructions to deploy it locally.

    • You can use any tools or language as long as the deployment information are simple to follow

    • Be sure to remove your API keys before submission.

    • We recommend tools such as docker-compose for participants to easily and safely judge your submission.

    • We also recommend using either a public-weight models, or AI tools that have enough of a free trial for participants to test your submission

  • a brief report summarizing your project following this template

  • a 2-minute video demonstration of how your interactive demo works

  • Optionally, an endpoint for participants to test your application directly (for example through Vercel).

The submission will be reviewed under the following criteria:

  • Viscerality: Does the demo give you a felt sense of AI capabilities? Is it fun, engaging, immersive or understandable? Does it make you feel the real-world implications?

  • Importance: How important are the insights delivered by this demo for understanding how AI will affect humanity in the coming years?

  • Accuracy: If the demo shows capabilities or risks, does it present them accurately, giving takeaways that generalize well? If the demonstration explains a concept, is it accurate?

  • Ease of testing: How easy is it to try out the demo?

  • Safety: Is the demo safe enough to publish publicly without causing harm?

  • Overall: How good is the demo overall?

The judging panel's scores will count for half of the final score, with the other half coming from peer reviews. You will be assigned as a peer reviewer of some other hackathon submissions - you must review these to be eligible to win.

Top teams will win a share of our $2,000 prize pool:

🥇 1st place: $1,000

🥈 2nd place: $600

🥉 3rd place: $300

🎖️ 4th place: $100

Additionally, high quality submissions that are a good fit for AI Digest might be invited to collaborate to polish up and publish their demo there, to reach a wider audience of policymakers and the public.

Why should I join?

There’s loads of reasons to join! Here are just a few:

  • Participate in a game jam-style event focused on an AI safety-adjacent topic

  • Get to know new people interested in AI-safety

  • Win up to $1,000

  • Receive a certificate of participation

  • Contribute to important public outreach on AI safety

  • And many many more… Come along!

Do I need experience in AI safety to join?

Not at all! This can be an occasion for you to learn more about AI safety and public outreach. We provide code templates and ideas to kickstart your projects, mentors to give feedback on your project, and a great community of interested developers to give reviews and feedback on your project.

What if my project seems too risky to share?

Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.

For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy

I have more questions to ask

Head over to the Frequently Asked Questions on the submission tab, and feel free to ping us ( @mentor ) on Discord!

Resources

Resources

Arrow
Arrow
Arrow

Resources

  • Examples of AI tools you can use in your demo

    • Langchain: Library to call LLMs and has a universal interface, allowing you to easily switch to other models.

    • LLama3.1: A performant open-weight LLM model. We recommend using open-weight models for participant to more easily judge your demo.

    • Claude Sonnet 3.5: A private LLM model that is cheap and performant

    • GPT-4o: Private and performant LLM model

    • Retell.ai: A service for automated call platforms

    • Openrouter: Universal interface for LLM through API call

Schedule

Schedule

Arrow
Arrow
Arrow

The schedule runs from 4 PM UTC Friday to 3 AM Monday. We start with an introductory talk and end the event during the following week with an awards ceremony. Join the public ICal here.

You will also find Explorer events, such as collaborative brainstorming and team match-making before the hackathon begins on Discord and in the calendar.

Entries

Speakers & Collaborators

Lucas Hansen

Keynote Speaker & Judge

Lucas brings 10 years of technology experience to CivAI. Previously, he founded Qualia, a real estate software company which serves millions of homeowners each year.

Épiphanie Gédéon

Organizer, Mentor & Judge

Épiphanie is a member of Centre pour la Sécurité de l'IA and currently an Apart Research Fellow

Adam Binksmith

Organizer, Mentor & Judge

Adam runs Sage, an organization helping impactful communities figure out what's true.

Archana Vaidheeswaran

Organizer

Archana is responsible for organizing the Apart Sprints, research hackathons to solve the most important questions in AI safety.

Misha Yagudin

Mentor & Judge

Misha runs Arb Research and Samotsvety Forecasting. Misha currently advises at Sage

Charbel-Raphaël Ségerie

Judge

Charbel-Raphaël is the executive director at Centre pour la Sécurité de l'IA and has led ML4Good, a bootcamp focused on AI safety.

Natalia Pérez-Campanero Antolín

Judge

A research manager at Apart, Natalia has a PhD in Interdisciplinary Biosciences from Oxford and has run the Royal Society's Entrepreneur-in-Residence program.

Jason Schreiber

Organizer and Judge

Jason is co-director of Apart Research and leads Apart Lab, our remote-first AI safety research fellowship.

Esben Kran

Organizer

Esben is the co-director of Apart Research and specializes in organizing research teams on pivotal AI security questions.

Speakers & Collaborators

Lucas Hansen

Keynote Speaker & Judge

Lucas brings 10 years of technology experience to CivAI. Previously, he founded Qualia, a real estate software company which serves millions of homeowners each year.

Épiphanie Gédéon

Organizer, Mentor & Judge

Épiphanie is a member of Centre pour la Sécurité de l'IA and currently an Apart Research Fellow

Adam Binksmith

Organizer, Mentor & Judge

Adam runs Sage, an organization helping impactful communities figure out what's true.

Archana Vaidheeswaran

Organizer

Archana is responsible for organizing the Apart Sprints, research hackathons to solve the most important questions in AI safety.

Misha Yagudin

Mentor & Judge

Misha runs Arb Research and Samotsvety Forecasting. Misha currently advises at Sage

Charbel-Raphaël Ségerie

Judge

Charbel-Raphaël is the executive director at Centre pour la Sécurité de l'IA and has led ML4Good, a bootcamp focused on AI safety.

Natalia Pérez-Campanero Antolín

Judge

A research manager at Apart, Natalia has a PhD in Interdisciplinary Biosciences from Oxford and has run the Royal Society's Entrepreneur-in-Residence program.

Jason Schreiber

Organizer and Judge

Jason is co-director of Apart Research and leads Apart Lab, our remote-first AI safety research fellowship.

Esben Kran

Organizer

Esben is the co-director of Apart Research and specializes in organizing research teams on pivotal AI security questions.

Registered Jam Sites

Register A Location

Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.

The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.

Registered Jam Sites

Register A Location

Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.

The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.