apart sprints

Develop breakthrough ideas

Join our monthly hackathons and collaborate with brilliant minds worldwide on impactful AI safety research

Sprint Features

Arrow
Arrow
Arrow

In-Person & Online

Join events on the Discord or at our in-person locations around the world! Follow the calendar here.

Live Mentorship Q&A

Our expert team will be available to help with any questions and theory on the hackathon Discord.

For Everyone

You can join in the middle of the Sprint if you don't find time and we provide code starters, ideas and inspiration; see an example.

Next Steps

We will help you realize the impact of your research with the Apart Lab Fellowship, providing mentorship, help with publication, funding, and more.

Sprint Features

Arrow
Arrow

In-Person & Online

Join events on the Discord or at our in-person locations around the world! Follow the calendar here.

Live Mentorship Q&A

Our expert team will be available to help with any questions and theory on the hackathon Discord.

For Everyone

You can join in the middle of the Sprint if you don't find time and we provide code starters, ideas and inspiration; see an example.

Next Steps

We will help you realize the impact of your research with the Apart Lab Fellowship, providing mentorship, help with publication, funding, and more.

With partners and collaborators from

  • OpenAI logo
  • OpenAI logo
  • OpenAI logo
  • OpenAI logo

Recent Winning Hackathon Projects

Mar 19, 2025

Sandbag Detection through Model Degradation

We propose a novel technique to detect sandbagging in LLMs by adding varying amount of noise to model weights and monitoring performance.

Read More

Read More

Read More

Mar 19, 2025

AI Alignment Knowledge Graph

We present a web based interactive knowledge graph with concise topical summaries in the field of AI alignement

Read More

Read More

Read More

Mar 19, 2025

Speculative Consequences of A.I. Misuse

This project uses A.I. Technology to spoof an influential online figure, Mr Beast, and use him to promote a fake scam website we created.

Read More

Read More

Read More

Mar 19, 2025

DarkForest - Defending the Authentic and Humane Web

DarkForest is a pioneering Human Content Verification System (HCVS) designed to safeguard the authenticity of online spaces in the face of increasing AI-generated content. By leveraging graph-based reinforcement learning and blockchain technology, DarkForest proposes a novel approach to safeguarding the authentic and humane web. We aim to become the vanguard in the arms race between AI-generated content and human-centric online spaces.

Read More

Read More

Read More

Mar 19, 2025

Diamonds are Not All You Need

This project tests an AI agent in a straightforward alignment problem. The agent is given creative freedom within a Minecraft world and is tasked with transforming a 100x100 radius of the world into diamond. It is explicitly asked not to act outside the designated area. The AI agent can execute build commands and is regulated by a Safety System that comprises an oversight agent. The objective of this study is to observe the behavior of the AI agent in a sandboxed environment, record metrics on how effectively it accomplishes its task, how frequently it attempts unsafe behavior, and how it behaves in response to real-world feedback.

Read More

Read More

Read More

Mar 18, 2025

Robust Machine Unlearning for Dangerous Capabilities

We test different unlearning methods to make models more robust against exploitation by malicious actors for the creation of bioweapons.

Read More

Read More

Read More

Publications From Hackathons

Feb 20, 2025

Deception Detection Hackathon: Preventing AI deception

Read More

Read More

Read More

Mar 18, 2025

Safe ai

The rapid adoption of AI in critical industries like healthcare and legal services has highlighted the urgent need for robust risk mitigation mechanisms. While domain-specific AI agents offer efficiency, they often lack transparency and accountability, raising concerns about safety, reliability, and compliance. The stakes are high, as AI failures in these sectors can lead to catastrophic outcomes, including loss of life, legal repercussions, and significant financial and reputational damage. Current solutions, such as regulatory frameworks and quality assurance protocols, provide only partial protection against the multifaceted risks associated with AI deployment. This situation underscores the necessity for an innovative approach that combines comprehensive risk assessment with financial safeguards to ensure the responsible and secure implementation of AI technologies across high-stakes industries.

Read More

Read More

Read More

Mar 18, 2025

Cite2Root

Regain information autonomy by bringing people closer to the source of truth.

Read More

Read More

Read More

Mar 18, 2025

VaultX - AI-Driven Middleware for Real-Time PII Detection and Data Security

VaultX is an AI-powered middleware solution designed for real-time detection, encryption, and secure management of Personally Identifiable Information (PII). By integrating regex, NER, and Language Models, VaultX ensures accuracy and scalability, seamlessly integrating into workflows like chatbots, web forms, and document processing. It helps businesses comply with global data privacy laws while safeguarding sensitive data from breaches and misuse.

Read More

Read More

Read More

Mar 18, 2025

.ALign File

In a post-AGI future, misaligned AI systems risk harmful consequences, especially with control over critical infrastructure. The Alignment Compliance Framework (ACF) ensures ethical AI adherence using .align files, Alignment Testing, and Decentralized Identifiers (DIDs). This scalable, decentralized system integrates alignment into development and lifecycle monitoring. ACF offers secure libraries, no-code tools for AI creation, regulatory compliance, continuous monitoring, and advisory services, promoting safer, commercially viable AI deployment.

Read More

Read More

Read More

Mar 18, 2025

Scoped LLM: Enhancing Adversarial Robustness and Security Through Targeted Model Scoping

Even with Reinforcement Learning from Human or AI Feedback (RLHF/RLAIF)

to avoid harmful outputs, fine-tuned Large Language Models (LLMs) often present insufficient

refusals due to adversarial attacks causing them to revert to reveal harmful knowledge from

pre-training. Machine unlearning has emerged as an alternative, aiming to remove harmful

knowledge permanently, but it relies on explicitly anticipating threats, leaving models exposed

to unforeseen risks. This project introduces model scoping, a novel approach to apply a least

privilege mindset to LLM safety and limit interactions to a predefined domain. By narrowing

the model’s operational domain, model scoping reduces susceptibility to adversarial prompts

and unforeseen misuse. This strategy offers a more robust framework for safe AI deployment in

unpredictable, evolving environments.

Read More

Read More

Read More

Mar 29, 2025

-

Mar 30, 2025

London & Online

AI Control Hackathon 2025

This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible

Learn More

Learn More

Learn More

Learn More

Mar 7, 2025

-

Mar 14, 2025

Online & In-person

Women in AI Safety Hackathon

This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible

Learn More

Learn More

Learn More

Learn More

Jan 17, 2025

-

Jan 20, 2025

Online & In-Person

AI Safety Entrepreneurship Hackathon

This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible

Learn More

Learn More

Learn More

Learn More

Nov 23, 2024

-

Nov 25, 2024

Online & In-Person

Autostructures: interfaces not between humans and AI, but between humans *via* AI

This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible

Learn More

Learn More

Learn More

Learn More

Nov 22, 2024

-

Nov 25, 2024

Online & In-Person

Reprogramming AI Models Hackathon

This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible

Learn More

Learn More

Learn More

Learn More

Nov 21, 2024

-

Nov 22, 2024

Washington, D.C. & Online

Howard University AI Safety Summit & Policy Hackathon

This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible

Learn More

Learn More

Learn More

Learn More