Jun 28, 2024
-
Jul 1, 2024
Online & In-Person
Deception Detection Hackathon: Preventing AI deception




re you fascinated by the incredible advancements in AI? As AI becomes smarter and more powerful, it's essential that we make sure it always tells the truth and doesn't trick people. Imagine if an AI could manipulate narratives or try to scheme—that could lead to serious problems!
00:00:00:00
00:00:00:00
00:00:00:00
00:00:00:00
This event is ongoing.
This event has concluded.
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

Are you fascinated by the incredible advancements in AI? As AI becomes smarter and more powerful, it's essential that we make sure it always tells the truth and doesn't trick people. Imagine if an AI could manipulate narratives or try to scheme—that could lead to serious problems!
That's where you come in. We're inviting you to join the Deception Detection Hackathon, an exciting event where you'll team up with researchers, programmers, and AI safety experts to create amazing pilot experiments to spot when an AI is being deceptive (and potentially reducing deceptive tendencies!). Over one thrilling weekend, you'll put your skills to the test and develop cutting-edge techniques to keep AI honest and trustworthy.
Why deception detection matters
Deception in AI, a concept severely under-explored, occurs when an AI system is capable of deceiving a user, either designed for it by a malicious actor or due to misaligned goals. Such systems may appear to be aligned with users and humans values during training and evaluation but pursue malign objectives when deployed, potentially causing harm or undermining trust in AI.
Examples of such work can be found in:
Situational Awareness Benchmark: An evaluation of how well models can identify what state they are in (hackathon project, talk)
LLM Identification Benchmark: Can an LLM identify whether a human or an LLM wrote a text? Might be important for self-coordination across chat instances (similar hackathon project)
Strategic Deception Report: An example of strategic deception presented by Apollo Research at the AI Safety Summit 2023 (explainer)
To mitigate these risks, we must develop robust deception detection methods that can identify instances of strategic deception, make headway on understanding AI capabilities for deception, and prevent AI systems from misleading humans. By participating in this hackathon, you'll contribute to the critical task of ensuring that AI remains transparent, accountable, and aligned with human values.
Contribute to AGI deception research

During the hackathon, you'll have the opportunity to:
Learn from experts in AI safety, deceptive alignment, and strategic deception
Collaborate with a diverse group of participants to ideate and develop deception detection techniques
Create benchmarks and evaluation methods to assess the effectiveness of deception detection approaches
Compete for prizes and recognition for the most innovative and impactful solutions
Network with like-minded individuals passionate about ensuring the safety and trustworthiness of AI
Whether you're an AI researcher, developer, or enthusiast, this hackathon provides a unique platform to apply your skills and knowledge to address one of the most pressing challenges in AI safety.
Join us in late June for a weekend of collaboration, innovation, and problem-solving as we work together to prevent AI from deceiving humans. Stay tuned for more details on the exact dates, format, and registration process.
Don't miss this opportunity to contribute to the development of trustworthy AI systems and help shape a future where AI and humans can work together safely and transparently. Let's hack for a deception-free AI future!
Prizes, evaluation, and submission
You will join in teams to submit a PDF about your research according to the submission template shared in the submission tab! Depending on the judge's reviews, you'll have the chance to win from the $2,000 prize pool! Find the review criteria on the submission tab.
🥇 $1,000 for the top team
🥈 $600 for the second prize
🥉 $300 for the third prize
🏅 $100 for the fourth prize
What is a research hackathon?
The AGI Deception Detection Hackathon is a weekend-long event where you participate in teams (1-5) to create interesting, fun, and impactful research. You submit a PDF report that summarizes and discusses your findings in the context of AI safety. These reports will be judged by our panel and you can win up to $1,000!
It runs from 28th June to 1st July and we're excited to welcome you for a weekend of engaging research. You will hear fascinating talks about real-world projects tackling these types of questions, get the opportunity to discuss your ideas with experienced mentors, and you will get reviews from top-tier researchers in the field of AI safety to further your exploration.
Everyone can participate and we encourage you to join especially if you’re considering AI safety from another career. We give you code templates and ideas to kickstart your projects and you’ll be surprised what you can accomplish in just a weekend – especially with your new-found community!
Read more about what you can expect, the schedule, and what previous participants have said about being part of the hackathon below.
Why should I join?
There’s loads of reasons to join! Here are just a few:
See how fun and interesting AI safety can be
Get to know new people who are into the overlap of empirical ML safety and AI governance
Win up to $1,000, helping you towards your first H100 GPU
Get practical experience with LLM evaluations and AI safety research
Show the AI safety labs what you are able to do to increase your chances at some amazing jobs
Get a certificate at the end!
Get proof that your work is awesome so you can get that grant to pursue the AI safety research that you always wanted to pursue
The best teams are offered to participate in the Apart Lab program, which supports teams in their journey towards publishing groundbreaking AI safety and security research
And many many more… Come along!
Do I need experience in AI safety to join?
Please join! This can be your first foray into AI and ML safety and maybe you’ll realize that there are exciting low-hanging fruit that your specific skillset is adapted to. Even if you normally don’t find it particularly interesting, this time you might see it in a new light!
There’s a lot of pressure from AI safety to perform at a top level and this seems to drive some people out of the field. We’d love it if you consider joining with a mindset of fun exploration and get a positive experience out of the weekend.
What are previous experiences from the research hackathon?
Yoann Poupart, BlockLoads CTO: "This Hackathon was a perfect blend of learning, testing, and collaboration on cutting-edge AI Safety research. I really feel that I gained practical knowledge that cannot be learned only by reading articles.”
Lucie Philippon, France Pacific Territories Economic Committee: "It was great meeting such cool people to work with over the weekend! I did not know any of the other people in my group at first, and now I'm looking forward to working with them again on research projects! The organizers were also super helpful and contributed a lot to the success of our project.”
Akash Kundu, now an Apart Lab fellow: "It was an amazing experience working with people I didn't even know before the hackathon. All three of my teammates were extremely spread out, while I am from India, my teammates were from New York and Taiwan. It was amazing how we pulled this off in 48 hours in spite of the time difference. Moreover, the mentors were extremely encouraging and supportive which helped us gain clarity whenever we got stuck and helped us create an interesting project in the end.”
Nora Petrova, ML Engineer at Prolific: “The hackathon really helped me to be embedded in a community where everyone was working on the same topic. There was a lot of curiosity and interest in the community. Getting feedback from others was interesting as well and I could see how other researchers perceived my project. It was also really interesting to see all the other projects and it was positive to see other's work on it.”
Chris Mathwin, MATS Scholar: "The Interpretability Hackathon exceeded my expectations, it was incredibly well organized with an intelligently curated list of very helpful resources. I had a lot of fun participating and genuinely feel I was able to learn significantly more than I would have, had I spent my time elsewhere. I highly recommend these events to anyone who is interested in this sort of work!”
What if my research seems too risky to share?
Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.
For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy.
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

Are you fascinated by the incredible advancements in AI? As AI becomes smarter and more powerful, it's essential that we make sure it always tells the truth and doesn't trick people. Imagine if an AI could manipulate narratives or try to scheme—that could lead to serious problems!
That's where you come in. We're inviting you to join the Deception Detection Hackathon, an exciting event where you'll team up with researchers, programmers, and AI safety experts to create amazing pilot experiments to spot when an AI is being deceptive (and potentially reducing deceptive tendencies!). Over one thrilling weekend, you'll put your skills to the test and develop cutting-edge techniques to keep AI honest and trustworthy.
Why deception detection matters
Deception in AI, a concept severely under-explored, occurs when an AI system is capable of deceiving a user, either designed for it by a malicious actor or due to misaligned goals. Such systems may appear to be aligned with users and humans values during training and evaluation but pursue malign objectives when deployed, potentially causing harm or undermining trust in AI.
Examples of such work can be found in:
Situational Awareness Benchmark: An evaluation of how well models can identify what state they are in (hackathon project, talk)
LLM Identification Benchmark: Can an LLM identify whether a human or an LLM wrote a text? Might be important for self-coordination across chat instances (similar hackathon project)
Strategic Deception Report: An example of strategic deception presented by Apollo Research at the AI Safety Summit 2023 (explainer)
To mitigate these risks, we must develop robust deception detection methods that can identify instances of strategic deception, make headway on understanding AI capabilities for deception, and prevent AI systems from misleading humans. By participating in this hackathon, you'll contribute to the critical task of ensuring that AI remains transparent, accountable, and aligned with human values.
Contribute to AGI deception research

During the hackathon, you'll have the opportunity to:
Learn from experts in AI safety, deceptive alignment, and strategic deception
Collaborate with a diverse group of participants to ideate and develop deception detection techniques
Create benchmarks and evaluation methods to assess the effectiveness of deception detection approaches
Compete for prizes and recognition for the most innovative and impactful solutions
Network with like-minded individuals passionate about ensuring the safety and trustworthiness of AI
Whether you're an AI researcher, developer, or enthusiast, this hackathon provides a unique platform to apply your skills and knowledge to address one of the most pressing challenges in AI safety.
Join us in late June for a weekend of collaboration, innovation, and problem-solving as we work together to prevent AI from deceiving humans. Stay tuned for more details on the exact dates, format, and registration process.
Don't miss this opportunity to contribute to the development of trustworthy AI systems and help shape a future where AI and humans can work together safely and transparently. Let's hack for a deception-free AI future!
Prizes, evaluation, and submission
You will join in teams to submit a PDF about your research according to the submission template shared in the submission tab! Depending on the judge's reviews, you'll have the chance to win from the $2,000 prize pool! Find the review criteria on the submission tab.
🥇 $1,000 for the top team
🥈 $600 for the second prize
🥉 $300 for the third prize
🏅 $100 for the fourth prize
What is a research hackathon?
The AGI Deception Detection Hackathon is a weekend-long event where you participate in teams (1-5) to create interesting, fun, and impactful research. You submit a PDF report that summarizes and discusses your findings in the context of AI safety. These reports will be judged by our panel and you can win up to $1,000!
It runs from 28th June to 1st July and we're excited to welcome you for a weekend of engaging research. You will hear fascinating talks about real-world projects tackling these types of questions, get the opportunity to discuss your ideas with experienced mentors, and you will get reviews from top-tier researchers in the field of AI safety to further your exploration.
Everyone can participate and we encourage you to join especially if you’re considering AI safety from another career. We give you code templates and ideas to kickstart your projects and you’ll be surprised what you can accomplish in just a weekend – especially with your new-found community!
Read more about what you can expect, the schedule, and what previous participants have said about being part of the hackathon below.
Why should I join?
There’s loads of reasons to join! Here are just a few:
See how fun and interesting AI safety can be
Get to know new people who are into the overlap of empirical ML safety and AI governance
Win up to $1,000, helping you towards your first H100 GPU
Get practical experience with LLM evaluations and AI safety research
Show the AI safety labs what you are able to do to increase your chances at some amazing jobs
Get a certificate at the end!
Get proof that your work is awesome so you can get that grant to pursue the AI safety research that you always wanted to pursue
The best teams are offered to participate in the Apart Lab program, which supports teams in their journey towards publishing groundbreaking AI safety and security research
And many many more… Come along!
Do I need experience in AI safety to join?
Please join! This can be your first foray into AI and ML safety and maybe you’ll realize that there are exciting low-hanging fruit that your specific skillset is adapted to. Even if you normally don’t find it particularly interesting, this time you might see it in a new light!
There’s a lot of pressure from AI safety to perform at a top level and this seems to drive some people out of the field. We’d love it if you consider joining with a mindset of fun exploration and get a positive experience out of the weekend.
What are previous experiences from the research hackathon?
Yoann Poupart, BlockLoads CTO: "This Hackathon was a perfect blend of learning, testing, and collaboration on cutting-edge AI Safety research. I really feel that I gained practical knowledge that cannot be learned only by reading articles.”
Lucie Philippon, France Pacific Territories Economic Committee: "It was great meeting such cool people to work with over the weekend! I did not know any of the other people in my group at first, and now I'm looking forward to working with them again on research projects! The organizers were also super helpful and contributed a lot to the success of our project.”
Akash Kundu, now an Apart Lab fellow: "It was an amazing experience working with people I didn't even know before the hackathon. All three of my teammates were extremely spread out, while I am from India, my teammates were from New York and Taiwan. It was amazing how we pulled this off in 48 hours in spite of the time difference. Moreover, the mentors were extremely encouraging and supportive which helped us gain clarity whenever we got stuck and helped us create an interesting project in the end.”
Nora Petrova, ML Engineer at Prolific: “The hackathon really helped me to be embedded in a community where everyone was working on the same topic. There was a lot of curiosity and interest in the community. Getting feedback from others was interesting as well and I could see how other researchers perceived my project. It was also really interesting to see all the other projects and it was positive to see other's work on it.”
Chris Mathwin, MATS Scholar: "The Interpretability Hackathon exceeded my expectations, it was incredibly well organized with an intelligently curated list of very helpful resources. I had a lot of fun participating and genuinely feel I was able to learn significantly more than I would have, had I spent my time elsewhere. I highly recommend these events to anyone who is interested in this sort of work!”
What if my research seems too risky to share?
Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.
For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy.
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

Are you fascinated by the incredible advancements in AI? As AI becomes smarter and more powerful, it's essential that we make sure it always tells the truth and doesn't trick people. Imagine if an AI could manipulate narratives or try to scheme—that could lead to serious problems!
That's where you come in. We're inviting you to join the Deception Detection Hackathon, an exciting event where you'll team up with researchers, programmers, and AI safety experts to create amazing pilot experiments to spot when an AI is being deceptive (and potentially reducing deceptive tendencies!). Over one thrilling weekend, you'll put your skills to the test and develop cutting-edge techniques to keep AI honest and trustworthy.
Why deception detection matters
Deception in AI, a concept severely under-explored, occurs when an AI system is capable of deceiving a user, either designed for it by a malicious actor or due to misaligned goals. Such systems may appear to be aligned with users and humans values during training and evaluation but pursue malign objectives when deployed, potentially causing harm or undermining trust in AI.
Examples of such work can be found in:
Situational Awareness Benchmark: An evaluation of how well models can identify what state they are in (hackathon project, talk)
LLM Identification Benchmark: Can an LLM identify whether a human or an LLM wrote a text? Might be important for self-coordination across chat instances (similar hackathon project)
Strategic Deception Report: An example of strategic deception presented by Apollo Research at the AI Safety Summit 2023 (explainer)
To mitigate these risks, we must develop robust deception detection methods that can identify instances of strategic deception, make headway on understanding AI capabilities for deception, and prevent AI systems from misleading humans. By participating in this hackathon, you'll contribute to the critical task of ensuring that AI remains transparent, accountable, and aligned with human values.
Contribute to AGI deception research

During the hackathon, you'll have the opportunity to:
Learn from experts in AI safety, deceptive alignment, and strategic deception
Collaborate with a diverse group of participants to ideate and develop deception detection techniques
Create benchmarks and evaluation methods to assess the effectiveness of deception detection approaches
Compete for prizes and recognition for the most innovative and impactful solutions
Network with like-minded individuals passionate about ensuring the safety and trustworthiness of AI
Whether you're an AI researcher, developer, or enthusiast, this hackathon provides a unique platform to apply your skills and knowledge to address one of the most pressing challenges in AI safety.
Join us in late June for a weekend of collaboration, innovation, and problem-solving as we work together to prevent AI from deceiving humans. Stay tuned for more details on the exact dates, format, and registration process.
Don't miss this opportunity to contribute to the development of trustworthy AI systems and help shape a future where AI and humans can work together safely and transparently. Let's hack for a deception-free AI future!
Prizes, evaluation, and submission
You will join in teams to submit a PDF about your research according to the submission template shared in the submission tab! Depending on the judge's reviews, you'll have the chance to win from the $2,000 prize pool! Find the review criteria on the submission tab.
🥇 $1,000 for the top team
🥈 $600 for the second prize
🥉 $300 for the third prize
🏅 $100 for the fourth prize
What is a research hackathon?
The AGI Deception Detection Hackathon is a weekend-long event where you participate in teams (1-5) to create interesting, fun, and impactful research. You submit a PDF report that summarizes and discusses your findings in the context of AI safety. These reports will be judged by our panel and you can win up to $1,000!
It runs from 28th June to 1st July and we're excited to welcome you for a weekend of engaging research. You will hear fascinating talks about real-world projects tackling these types of questions, get the opportunity to discuss your ideas with experienced mentors, and you will get reviews from top-tier researchers in the field of AI safety to further your exploration.
Everyone can participate and we encourage you to join especially if you’re considering AI safety from another career. We give you code templates and ideas to kickstart your projects and you’ll be surprised what you can accomplish in just a weekend – especially with your new-found community!
Read more about what you can expect, the schedule, and what previous participants have said about being part of the hackathon below.
Why should I join?
There’s loads of reasons to join! Here are just a few:
See how fun and interesting AI safety can be
Get to know new people who are into the overlap of empirical ML safety and AI governance
Win up to $1,000, helping you towards your first H100 GPU
Get practical experience with LLM evaluations and AI safety research
Show the AI safety labs what you are able to do to increase your chances at some amazing jobs
Get a certificate at the end!
Get proof that your work is awesome so you can get that grant to pursue the AI safety research that you always wanted to pursue
The best teams are offered to participate in the Apart Lab program, which supports teams in their journey towards publishing groundbreaking AI safety and security research
And many many more… Come along!
Do I need experience in AI safety to join?
Please join! This can be your first foray into AI and ML safety and maybe you’ll realize that there are exciting low-hanging fruit that your specific skillset is adapted to. Even if you normally don’t find it particularly interesting, this time you might see it in a new light!
There’s a lot of pressure from AI safety to perform at a top level and this seems to drive some people out of the field. We’d love it if you consider joining with a mindset of fun exploration and get a positive experience out of the weekend.
What are previous experiences from the research hackathon?
Yoann Poupart, BlockLoads CTO: "This Hackathon was a perfect blend of learning, testing, and collaboration on cutting-edge AI Safety research. I really feel that I gained practical knowledge that cannot be learned only by reading articles.”
Lucie Philippon, France Pacific Territories Economic Committee: "It was great meeting such cool people to work with over the weekend! I did not know any of the other people in my group at first, and now I'm looking forward to working with them again on research projects! The organizers were also super helpful and contributed a lot to the success of our project.”
Akash Kundu, now an Apart Lab fellow: "It was an amazing experience working with people I didn't even know before the hackathon. All three of my teammates were extremely spread out, while I am from India, my teammates were from New York and Taiwan. It was amazing how we pulled this off in 48 hours in spite of the time difference. Moreover, the mentors were extremely encouraging and supportive which helped us gain clarity whenever we got stuck and helped us create an interesting project in the end.”
Nora Petrova, ML Engineer at Prolific: “The hackathon really helped me to be embedded in a community where everyone was working on the same topic. There was a lot of curiosity and interest in the community. Getting feedback from others was interesting as well and I could see how other researchers perceived my project. It was also really interesting to see all the other projects and it was positive to see other's work on it.”
Chris Mathwin, MATS Scholar: "The Interpretability Hackathon exceeded my expectations, it was incredibly well organized with an intelligently curated list of very helpful resources. I had a lot of fun participating and genuinely feel I was able to learn significantly more than I would have, had I spent my time elsewhere. I highly recommend these events to anyone who is interested in this sort of work!”
What if my research seems too risky to share?
Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.
For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy.
Sign Ups
Entries
Overview
Resources
Schedule
Entries
Overview

Are you fascinated by the incredible advancements in AI? As AI becomes smarter and more powerful, it's essential that we make sure it always tells the truth and doesn't trick people. Imagine if an AI could manipulate narratives or try to scheme—that could lead to serious problems!
That's where you come in. We're inviting you to join the Deception Detection Hackathon, an exciting event where you'll team up with researchers, programmers, and AI safety experts to create amazing pilot experiments to spot when an AI is being deceptive (and potentially reducing deceptive tendencies!). Over one thrilling weekend, you'll put your skills to the test and develop cutting-edge techniques to keep AI honest and trustworthy.
Why deception detection matters
Deception in AI, a concept severely under-explored, occurs when an AI system is capable of deceiving a user, either designed for it by a malicious actor or due to misaligned goals. Such systems may appear to be aligned with users and humans values during training and evaluation but pursue malign objectives when deployed, potentially causing harm or undermining trust in AI.
Examples of such work can be found in:
Situational Awareness Benchmark: An evaluation of how well models can identify what state they are in (hackathon project, talk)
LLM Identification Benchmark: Can an LLM identify whether a human or an LLM wrote a text? Might be important for self-coordination across chat instances (similar hackathon project)
Strategic Deception Report: An example of strategic deception presented by Apollo Research at the AI Safety Summit 2023 (explainer)
To mitigate these risks, we must develop robust deception detection methods that can identify instances of strategic deception, make headway on understanding AI capabilities for deception, and prevent AI systems from misleading humans. By participating in this hackathon, you'll contribute to the critical task of ensuring that AI remains transparent, accountable, and aligned with human values.
Contribute to AGI deception research

During the hackathon, you'll have the opportunity to:
Learn from experts in AI safety, deceptive alignment, and strategic deception
Collaborate with a diverse group of participants to ideate and develop deception detection techniques
Create benchmarks and evaluation methods to assess the effectiveness of deception detection approaches
Compete for prizes and recognition for the most innovative and impactful solutions
Network with like-minded individuals passionate about ensuring the safety and trustworthiness of AI
Whether you're an AI researcher, developer, or enthusiast, this hackathon provides a unique platform to apply your skills and knowledge to address one of the most pressing challenges in AI safety.
Join us in late June for a weekend of collaboration, innovation, and problem-solving as we work together to prevent AI from deceiving humans. Stay tuned for more details on the exact dates, format, and registration process.
Don't miss this opportunity to contribute to the development of trustworthy AI systems and help shape a future where AI and humans can work together safely and transparently. Let's hack for a deception-free AI future!
Prizes, evaluation, and submission
You will join in teams to submit a PDF about your research according to the submission template shared in the submission tab! Depending on the judge's reviews, you'll have the chance to win from the $2,000 prize pool! Find the review criteria on the submission tab.
🥇 $1,000 for the top team
🥈 $600 for the second prize
🥉 $300 for the third prize
🏅 $100 for the fourth prize
What is a research hackathon?
The AGI Deception Detection Hackathon is a weekend-long event where you participate in teams (1-5) to create interesting, fun, and impactful research. You submit a PDF report that summarizes and discusses your findings in the context of AI safety. These reports will be judged by our panel and you can win up to $1,000!
It runs from 28th June to 1st July and we're excited to welcome you for a weekend of engaging research. You will hear fascinating talks about real-world projects tackling these types of questions, get the opportunity to discuss your ideas with experienced mentors, and you will get reviews from top-tier researchers in the field of AI safety to further your exploration.
Everyone can participate and we encourage you to join especially if you’re considering AI safety from another career. We give you code templates and ideas to kickstart your projects and you’ll be surprised what you can accomplish in just a weekend – especially with your new-found community!
Read more about what you can expect, the schedule, and what previous participants have said about being part of the hackathon below.
Why should I join?
There’s loads of reasons to join! Here are just a few:
See how fun and interesting AI safety can be
Get to know new people who are into the overlap of empirical ML safety and AI governance
Win up to $1,000, helping you towards your first H100 GPU
Get practical experience with LLM evaluations and AI safety research
Show the AI safety labs what you are able to do to increase your chances at some amazing jobs
Get a certificate at the end!
Get proof that your work is awesome so you can get that grant to pursue the AI safety research that you always wanted to pursue
The best teams are offered to participate in the Apart Lab program, which supports teams in their journey towards publishing groundbreaking AI safety and security research
And many many more… Come along!
Do I need experience in AI safety to join?
Please join! This can be your first foray into AI and ML safety and maybe you’ll realize that there are exciting low-hanging fruit that your specific skillset is adapted to. Even if you normally don’t find it particularly interesting, this time you might see it in a new light!
There’s a lot of pressure from AI safety to perform at a top level and this seems to drive some people out of the field. We’d love it if you consider joining with a mindset of fun exploration and get a positive experience out of the weekend.
What are previous experiences from the research hackathon?
Yoann Poupart, BlockLoads CTO: "This Hackathon was a perfect blend of learning, testing, and collaboration on cutting-edge AI Safety research. I really feel that I gained practical knowledge that cannot be learned only by reading articles.”
Lucie Philippon, France Pacific Territories Economic Committee: "It was great meeting such cool people to work with over the weekend! I did not know any of the other people in my group at first, and now I'm looking forward to working with them again on research projects! The organizers were also super helpful and contributed a lot to the success of our project.”
Akash Kundu, now an Apart Lab fellow: "It was an amazing experience working with people I didn't even know before the hackathon. All three of my teammates were extremely spread out, while I am from India, my teammates were from New York and Taiwan. It was amazing how we pulled this off in 48 hours in spite of the time difference. Moreover, the mentors were extremely encouraging and supportive which helped us gain clarity whenever we got stuck and helped us create an interesting project in the end.”
Nora Petrova, ML Engineer at Prolific: “The hackathon really helped me to be embedded in a community where everyone was working on the same topic. There was a lot of curiosity and interest in the community. Getting feedback from others was interesting as well and I could see how other researchers perceived my project. It was also really interesting to see all the other projects and it was positive to see other's work on it.”
Chris Mathwin, MATS Scholar: "The Interpretability Hackathon exceeded my expectations, it was incredibly well organized with an intelligently curated list of very helpful resources. I had a lot of fun participating and genuinely feel I was able to learn significantly more than I would have, had I spent my time elsewhere. I highly recommend these events to anyone who is interested in this sort of work!”
What if my research seems too risky to share?
Besides emphasizing the introduction of concrete mitigation ideas for the risks presented, we are aware that projects emerging from this hackathon might pose a risk if disseminated irresponsibly.
For all of Apart's research events and dissemination, we follow our Responsible Disclosure Policy.
Speakers & Collaborators
Marius Hobbhahn
Co-organizer
Marius is the CEO of Apollo Research, a research non-profit that specializes in creating evaluations for deception.
Archana Vaidheeswaran
Organizer
Archana is responsible for organizing the Apart Sprints, research hackathons to solve the most important questions in AI safety.
Esben Kran
Organizer
Esben is the co-director of Apart Research and specializes in organizing research teams on pivotal AI security questions.
Jason Schreiber
Organizer and Judge
Jason is co-director of Apart Research and leads Apart Lab, our remote-first AI safety research fellowship.
Rudolf Laine
Judge & Reviewer
Def/acc EF cohort member. Author of the situational awareness benchmark and an independent AI safety researcher working with Owaine Evans.
Jacob Haimes
Speaker & Mentor
Author of the unpublished retro-holdout paper about evaluation datasets that have leaked into the training set and a fellow at the Apart Lab. Hosts a podcast on AI safety.
Kunvar Thaman
Reviewer
RE in mechanistic interpretability and former cybersecurity engineer. Author of the "Benchmark Inflation" paper quantifying LLM performance gaps. Apart Research Fellow.
Henry Sleight
Reviewer
Henry is a scholar support specialist at the MATS program and has supported Ethan Perez' research assistants in AI safety research. Board member of LISA.
Natalia Pérez-Campanero Antolín
Judge
A research manager at Apart, Natalia has a PhD in Interdisciplinary Biosciences from Oxford and has run the Royal Society's Entrepreneur-in-Residence program.
Mikita Balesni
Speaker
Mikita is a research scientist at Apollo Research and works on evaluation of dangerous language model capabilities. He helped Ukraine's cyber front against the Russian invasion
Speakers & Collaborators

Marius Hobbhahn
Co-organizer
Marius is the CEO of Apollo Research, a research non-profit that specializes in creating evaluations for deception.

Archana Vaidheeswaran
Organizer
Archana is responsible for organizing the Apart Sprints, research hackathons to solve the most important questions in AI safety.

Esben Kran
Organizer
Esben is the co-director of Apart Research and specializes in organizing research teams on pivotal AI security questions.

Jason Schreiber
Organizer and Judge
Jason is co-director of Apart Research and leads Apart Lab, our remote-first AI safety research fellowship.

Rudolf Laine
Judge & Reviewer
Def/acc EF cohort member. Author of the situational awareness benchmark and an independent AI safety researcher working with Owaine Evans.

Jacob Haimes
Speaker & Mentor
Author of the unpublished retro-holdout paper about evaluation datasets that have leaked into the training set and a fellow at the Apart Lab. Hosts a podcast on AI safety.

Kunvar Thaman
Reviewer
RE in mechanistic interpretability and former cybersecurity engineer. Author of the "Benchmark Inflation" paper quantifying LLM performance gaps. Apart Research Fellow.

Henry Sleight
Reviewer
Henry is a scholar support specialist at the MATS program and has supported Ethan Perez' research assistants in AI safety research. Board member of LISA.

Natalia Pérez-Campanero Antolín
Judge
A research manager at Apart, Natalia has a PhD in Interdisciplinary Biosciences from Oxford and has run the Royal Society's Entrepreneur-in-Residence program.
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Our Other Sprints
Apr 25, 2025
-
Apr 27, 2025
Research
Economics of Transformative AI: Research Sprint
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up
Apr 25, 2025
-
Apr 26, 2025
Research
Berkeley AI Policy Hackathon
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events