Jul 25, 2025
-
Jul 27, 2025
Online
AI Safety x Physics Grand Challenge



Ready to bridge physics and AI safety? Apply your rigorous quantitative training to one of the most impactful technical challenges of our time.
00:00:00:00
Days To Go
00:00:00:00
Days To Go
00:00:00:00
Days To Go
00:00:00:00
Days To Go
Ready to bridge physics and AI safety? Apply your rigorous quantitative training to one of the most impactful technical challenges of our time.
This event is ongoing.
This event has concluded.
Sign Ups
Entries
Overview
Resources
Guidelines
Schedule
Entries
Overview

This Challenge has been concluded
Over the July 25th-27th weekend intensive, participants tackled critical challenges by applying physics methodologies—from thermodynamics and statistical mechanics to computational mechanics—to understand and control AI systems. The event featured both Project and Exploration tracks, encouraging participants to bridge theoretical physics insights with practical AI safety solutions through collaborative innovation.
Following are the prize winners:
First Prize: Constrained Belief Updates in Transformer Representations:
Brianna Grado-White
This work extended computational mechanics theory to analyze how transformers implement optimal prediction under constraints. It showed belief state geometry across layers and that attention patterns decay exponentially with distance. Despite complex challenges, it bridged physics-inspired theory and practical interpretability, helping understand AI reasoning.
Second Prize: Momentum-Point-Perplexity Mechanics in Large Language Models
by Lorenzo Tomaz, Thomas Jones
This work found that energy remains approximately conserved across 20 different models. Their theoretical insights led to Jacobian steering—a control method that improves output quality while respecting the underlying physics, offering both fundamental understanding and practical applications for AI safety.Third Prize: Local Learning Coefficients in Reinforcement Learning:
by Jeremias Ferrao, Ilija LichkovskiThis project applies Singular Learning Theory to reinforcement learning, demonstrating how Local Learning Coefficients (LLCs) detect training phase transitions. By tracking reward components directly rather than derived metrics, LLC spikes correlate with capability milestones. This scalable approach enhances the understanding of how AI systems learn complex behaviors, with significant safety implications.
AI agentic system epidemiology
by Valentina Schastlivaia, Aray Karjauv
Using Physics-Informed Neural Networks to solve the governing equations, they created a framework for monitoring AI system health and predicting intervention effectiveness—providing a novel lens for understanding systemic AI risks.
Please find the link to the presentation here: https://docs.google.com/presentation/d/1IWKW9gMtc2qjfQAdWlRcT0EnJzeuJX3h6XOZC2oPOT4/edit?usp=sharing
Uniting Physics to Solve AI Safety
Join us for this AI Safety x Physics Grand Challenge to use your valuable skillset on chosen problem spaces for physicists!
You'll work alongside peers to create an original research project during this two day sprint with the guidance of expert mentors in the unique intersection of physics and AI safety. Be ready for excitement, progress, and, last but not least, prizes!
Physics and AI Safety?
The complexity of neural networks, the topology of their latent space, and the dynamics of the systems that deploy AGI today is an ideal topic of study for any physicist. We are at the absolute frontier of understanding learning dynamics through fundamental physics and during this event, you will become an integral part of this field's future!
Research Tracks

Employ your ML / AI experience from interpretability, benchmarking, or training to develop an original project employing the novel tools that physicists have introduced to the field in the past years.
Examples include the projects we've defined, such as operationalizing the natural abstraction hypothesis for empirical tests of compression or predictability of learned abstractions.

Propose novel research agendas to tackle the most pressing AI challenges or explore a potential bridge between an AI safety problem and a physics concept.
An example could be a proposal, outlining a novel physics-based solution to a key AI safety problem, with supporting material drawn from prior work.
Join
And experience…
Expert mentorship from leaders in the field of physics and AI safety
Structured problem areas with starter materials and concrete research directions
Global community of highly talented participants who are ambitiously solving top challenges
Follow-up support for exceptional projects through the Apart Research programs
Optional pre-event orientation and brainstorming for those new to AI safety
Shape the Future!
Join us in applying your academic training to one of the most important technical challenges of our time. Connect with a growing community of researchers safeguarding AI systems for the benefit of humanity.
FAQ
What Background Knowledge Is Required?
This challenge welcomes participants with physics backgrounds who are interested in applying rigorous methodologies to AI safety. While familiarity with machine learning is helpful, we encourage participation from:
Physics PhD students, postdocs, and faculty across all subfields
Theoretical physicists interested in complex systems and statistical mechanics
Experimental physicists with experience in data analysis and modeling
Computational physicists familiar with simulation and numerical methods
Physicists working at the intersection of physics and computer science
Anyone passionate about applying physics thinking to beneficial AI development
How Are Teams Formed?
Teams will self-organize during the first day through structured activities and interest-matching sessions. You can also form teams in advance with other registered participants via our Discord channel: https://ais.pub/disc
Teams should ideally combine:
Different physics backgrounds (theory, experiment, computation)
Varied experience levels with AI/ML
Complementary skills in mathematics, programming, and research
What Are The Expected Outputs?
Teams should produce research outputs and reports that demonstrate progress toward applying physics methodologies to AI safety challenges. The most promising projects may evolve into:
Academic papers or technical reports
Continued research collaboration with physics-AI safety researchers
Integration into ongoing research programs at leading institutions
Follow-up funding opportunities for extended research
Theoretical frameworks for future empirical validation
Will Compute Resources Be Provided?
Yes! Lambda is providing $400 in computing credits to each participating team, giving you access to powerful cloud instances (including A100s) to bring your ideas to life.
What Challenge Track Should I Choose?
You must align your project with one of our two tracks:
Project Track: This is a typical hackathon submission based on the starter materials or an original idea. We expect most of these to provide incremental theoretical progress and/or empirical evidence for (or against) an existing idea.
Relevant for higher context participants (i.e. those with experience running experiments on NNs or with an existing physics-AI agenda they want to accelerate progress on).
Exploration Track: Part of the idea for this hackathon is to open up and test the limits of a ‘physics for AI safety’ opportunity space. We have thought of some, but likely many more that would be exciting. Instead of traditional ‘hacking’, this will be a written submission, for example:
A proposal, outlining a novel physics-based solution to a key AI safety problem, with supporting material drawn from prior work.
A distillation or literature review bridging a CS/ML/AI idea with one from physics, with clear reference to an AI safety problem area.
We think this track has the potential to brainstorm some truly innovative ideas, while also allowing lower-context participants to get involved. For example, a ‘papers’ participant or group could partner with a ‘project’ group, resulting in a small research gain and a better understanding of the bigger picture.
What Problem Area Should I Focus On?
Your project should address one or more of our five problem areas:
Problem Area 1: Theory/Practice Gap for AI Interpretability - Apply physics methodologies (renormalization, computational mechanics, statistical mechanics) to build theoretical foundations for mechanistic interpretability
Problem Area 2: Learning, Inductive Biases, and Generalization - We want to understand how architecture, initialization, data distribution, and training procedures give rise to inductive biases—and in turn, how these biases control what is learned, when generalization occurs, and why models may fail.
Problem Area 3: Mathematical Models of Data Structure - AI alignment faces the challenge of understanding an AI's internal world model to prevent harmful actions. Physics methods, like coarse-graining and renormalization, could provide mathematical models to describe AI's learned representations and improve our understanding of AI's abstraction, generalization, and world modeling.
Problem Area 4: Quantitative Bounds on AI Behavior - This challenge focuses on using physics-inspired uncertainty quantification to establish quantitative bounds on AI behavior, ensuring alignment even when systems develop unanticipated strategies. Your work is crucial for guaranteeing ethical and reliable AI.
Problem Area 5: Scaling Laws, Phase Transitions, and Emergence - Uncover the physical foundations of AI performance, universal limits, and how smooth scaling reconciles with emergent capabilities.
Sign Ups
Entries
Overview
Resources
Guidelines
Schedule
Entries
Overview

This Challenge has been concluded
Over the July 25th-27th weekend intensive, participants tackled critical challenges by applying physics methodologies—from thermodynamics and statistical mechanics to computational mechanics—to understand and control AI systems. The event featured both Project and Exploration tracks, encouraging participants to bridge theoretical physics insights with practical AI safety solutions through collaborative innovation.
Following are the prize winners:
First Prize: Constrained Belief Updates in Transformer Representations:
Brianna Grado-White
This work extended computational mechanics theory to analyze how transformers implement optimal prediction under constraints. It showed belief state geometry across layers and that attention patterns decay exponentially with distance. Despite complex challenges, it bridged physics-inspired theory and practical interpretability, helping understand AI reasoning.
Second Prize: Momentum-Point-Perplexity Mechanics in Large Language Models
by Lorenzo Tomaz, Thomas Jones
This work found that energy remains approximately conserved across 20 different models. Their theoretical insights led to Jacobian steering—a control method that improves output quality while respecting the underlying physics, offering both fundamental understanding and practical applications for AI safety.Third Prize: Local Learning Coefficients in Reinforcement Learning:
by Jeremias Ferrao, Ilija LichkovskiThis project applies Singular Learning Theory to reinforcement learning, demonstrating how Local Learning Coefficients (LLCs) detect training phase transitions. By tracking reward components directly rather than derived metrics, LLC spikes correlate with capability milestones. This scalable approach enhances the understanding of how AI systems learn complex behaviors, with significant safety implications.
AI agentic system epidemiology
by Valentina Schastlivaia, Aray Karjauv
Using Physics-Informed Neural Networks to solve the governing equations, they created a framework for monitoring AI system health and predicting intervention effectiveness—providing a novel lens for understanding systemic AI risks.
Please find the link to the presentation here: https://docs.google.com/presentation/d/1IWKW9gMtc2qjfQAdWlRcT0EnJzeuJX3h6XOZC2oPOT4/edit?usp=sharing
Uniting Physics to Solve AI Safety
Join us for this AI Safety x Physics Grand Challenge to use your valuable skillset on chosen problem spaces for physicists!
You'll work alongside peers to create an original research project during this two day sprint with the guidance of expert mentors in the unique intersection of physics and AI safety. Be ready for excitement, progress, and, last but not least, prizes!
Physics and AI Safety?
The complexity of neural networks, the topology of their latent space, and the dynamics of the systems that deploy AGI today is an ideal topic of study for any physicist. We are at the absolute frontier of understanding learning dynamics through fundamental physics and during this event, you will become an integral part of this field's future!
Research Tracks

Employ your ML / AI experience from interpretability, benchmarking, or training to develop an original project employing the novel tools that physicists have introduced to the field in the past years.
Examples include the projects we've defined, such as operationalizing the natural abstraction hypothesis for empirical tests of compression or predictability of learned abstractions.

Propose novel research agendas to tackle the most pressing AI challenges or explore a potential bridge between an AI safety problem and a physics concept.
An example could be a proposal, outlining a novel physics-based solution to a key AI safety problem, with supporting material drawn from prior work.
Join
And experience…
Expert mentorship from leaders in the field of physics and AI safety
Structured problem areas with starter materials and concrete research directions
Global community of highly talented participants who are ambitiously solving top challenges
Follow-up support for exceptional projects through the Apart Research programs
Optional pre-event orientation and brainstorming for those new to AI safety
Shape the Future!
Join us in applying your academic training to one of the most important technical challenges of our time. Connect with a growing community of researchers safeguarding AI systems for the benefit of humanity.
FAQ
What Background Knowledge Is Required?
This challenge welcomes participants with physics backgrounds who are interested in applying rigorous methodologies to AI safety. While familiarity with machine learning is helpful, we encourage participation from:
Physics PhD students, postdocs, and faculty across all subfields
Theoretical physicists interested in complex systems and statistical mechanics
Experimental physicists with experience in data analysis and modeling
Computational physicists familiar with simulation and numerical methods
Physicists working at the intersection of physics and computer science
Anyone passionate about applying physics thinking to beneficial AI development
How Are Teams Formed?
Teams will self-organize during the first day through structured activities and interest-matching sessions. You can also form teams in advance with other registered participants via our Discord channel: https://ais.pub/disc
Teams should ideally combine:
Different physics backgrounds (theory, experiment, computation)
Varied experience levels with AI/ML
Complementary skills in mathematics, programming, and research
What Are The Expected Outputs?
Teams should produce research outputs and reports that demonstrate progress toward applying physics methodologies to AI safety challenges. The most promising projects may evolve into:
Academic papers or technical reports
Continued research collaboration with physics-AI safety researchers
Integration into ongoing research programs at leading institutions
Follow-up funding opportunities for extended research
Theoretical frameworks for future empirical validation
Will Compute Resources Be Provided?
Yes! Lambda is providing $400 in computing credits to each participating team, giving you access to powerful cloud instances (including A100s) to bring your ideas to life.
What Challenge Track Should I Choose?
You must align your project with one of our two tracks:
Project Track: This is a typical hackathon submission based on the starter materials or an original idea. We expect most of these to provide incremental theoretical progress and/or empirical evidence for (or against) an existing idea.
Relevant for higher context participants (i.e. those with experience running experiments on NNs or with an existing physics-AI agenda they want to accelerate progress on).
Exploration Track: Part of the idea for this hackathon is to open up and test the limits of a ‘physics for AI safety’ opportunity space. We have thought of some, but likely many more that would be exciting. Instead of traditional ‘hacking’, this will be a written submission, for example:
A proposal, outlining a novel physics-based solution to a key AI safety problem, with supporting material drawn from prior work.
A distillation or literature review bridging a CS/ML/AI idea with one from physics, with clear reference to an AI safety problem area.
We think this track has the potential to brainstorm some truly innovative ideas, while also allowing lower-context participants to get involved. For example, a ‘papers’ participant or group could partner with a ‘project’ group, resulting in a small research gain and a better understanding of the bigger picture.
What Problem Area Should I Focus On?
Your project should address one or more of our five problem areas:
Problem Area 1: Theory/Practice Gap for AI Interpretability - Apply physics methodologies (renormalization, computational mechanics, statistical mechanics) to build theoretical foundations for mechanistic interpretability
Problem Area 2: Learning, Inductive Biases, and Generalization - We want to understand how architecture, initialization, data distribution, and training procedures give rise to inductive biases—and in turn, how these biases control what is learned, when generalization occurs, and why models may fail.
Problem Area 3: Mathematical Models of Data Structure - AI alignment faces the challenge of understanding an AI's internal world model to prevent harmful actions. Physics methods, like coarse-graining and renormalization, could provide mathematical models to describe AI's learned representations and improve our understanding of AI's abstraction, generalization, and world modeling.
Problem Area 4: Quantitative Bounds on AI Behavior - This challenge focuses on using physics-inspired uncertainty quantification to establish quantitative bounds on AI behavior, ensuring alignment even when systems develop unanticipated strategies. Your work is crucial for guaranteeing ethical and reliable AI.
Problem Area 5: Scaling Laws, Phase Transitions, and Emergence - Uncover the physical foundations of AI performance, universal limits, and how smooth scaling reconciles with emergent capabilities.
Sign Ups
Entries
Overview
Resources
Guidelines
Schedule
Entries
Overview

This Challenge has been concluded
Over the July 25th-27th weekend intensive, participants tackled critical challenges by applying physics methodologies—from thermodynamics and statistical mechanics to computational mechanics—to understand and control AI systems. The event featured both Project and Exploration tracks, encouraging participants to bridge theoretical physics insights with practical AI safety solutions through collaborative innovation.
Following are the prize winners:
First Prize: Constrained Belief Updates in Transformer Representations:
Brianna Grado-White
This work extended computational mechanics theory to analyze how transformers implement optimal prediction under constraints. It showed belief state geometry across layers and that attention patterns decay exponentially with distance. Despite complex challenges, it bridged physics-inspired theory and practical interpretability, helping understand AI reasoning.
Second Prize: Momentum-Point-Perplexity Mechanics in Large Language Models
by Lorenzo Tomaz, Thomas Jones
This work found that energy remains approximately conserved across 20 different models. Their theoretical insights led to Jacobian steering—a control method that improves output quality while respecting the underlying physics, offering both fundamental understanding and practical applications for AI safety.Third Prize: Local Learning Coefficients in Reinforcement Learning:
by Jeremias Ferrao, Ilija LichkovskiThis project applies Singular Learning Theory to reinforcement learning, demonstrating how Local Learning Coefficients (LLCs) detect training phase transitions. By tracking reward components directly rather than derived metrics, LLC spikes correlate with capability milestones. This scalable approach enhances the understanding of how AI systems learn complex behaviors, with significant safety implications.
AI agentic system epidemiology
by Valentina Schastlivaia, Aray Karjauv
Using Physics-Informed Neural Networks to solve the governing equations, they created a framework for monitoring AI system health and predicting intervention effectiveness—providing a novel lens for understanding systemic AI risks.
Please find the link to the presentation here: https://docs.google.com/presentation/d/1IWKW9gMtc2qjfQAdWlRcT0EnJzeuJX3h6XOZC2oPOT4/edit?usp=sharing
Uniting Physics to Solve AI Safety
Join us for this AI Safety x Physics Grand Challenge to use your valuable skillset on chosen problem spaces for physicists!
You'll work alongside peers to create an original research project during this two day sprint with the guidance of expert mentors in the unique intersection of physics and AI safety. Be ready for excitement, progress, and, last but not least, prizes!
Physics and AI Safety?
The complexity of neural networks, the topology of their latent space, and the dynamics of the systems that deploy AGI today is an ideal topic of study for any physicist. We are at the absolute frontier of understanding learning dynamics through fundamental physics and during this event, you will become an integral part of this field's future!
Research Tracks

Employ your ML / AI experience from interpretability, benchmarking, or training to develop an original project employing the novel tools that physicists have introduced to the field in the past years.
Examples include the projects we've defined, such as operationalizing the natural abstraction hypothesis for empirical tests of compression or predictability of learned abstractions.

Propose novel research agendas to tackle the most pressing AI challenges or explore a potential bridge between an AI safety problem and a physics concept.
An example could be a proposal, outlining a novel physics-based solution to a key AI safety problem, with supporting material drawn from prior work.
Join
And experience…
Expert mentorship from leaders in the field of physics and AI safety
Structured problem areas with starter materials and concrete research directions
Global community of highly talented participants who are ambitiously solving top challenges
Follow-up support for exceptional projects through the Apart Research programs
Optional pre-event orientation and brainstorming for those new to AI safety
Shape the Future!
Join us in applying your academic training to one of the most important technical challenges of our time. Connect with a growing community of researchers safeguarding AI systems for the benefit of humanity.
FAQ
What Background Knowledge Is Required?
This challenge welcomes participants with physics backgrounds who are interested in applying rigorous methodologies to AI safety. While familiarity with machine learning is helpful, we encourage participation from:
Physics PhD students, postdocs, and faculty across all subfields
Theoretical physicists interested in complex systems and statistical mechanics
Experimental physicists with experience in data analysis and modeling
Computational physicists familiar with simulation and numerical methods
Physicists working at the intersection of physics and computer science
Anyone passionate about applying physics thinking to beneficial AI development
How Are Teams Formed?
Teams will self-organize during the first day through structured activities and interest-matching sessions. You can also form teams in advance with other registered participants via our Discord channel: https://ais.pub/disc
Teams should ideally combine:
Different physics backgrounds (theory, experiment, computation)
Varied experience levels with AI/ML
Complementary skills in mathematics, programming, and research
What Are The Expected Outputs?
Teams should produce research outputs and reports that demonstrate progress toward applying physics methodologies to AI safety challenges. The most promising projects may evolve into:
Academic papers or technical reports
Continued research collaboration with physics-AI safety researchers
Integration into ongoing research programs at leading institutions
Follow-up funding opportunities for extended research
Theoretical frameworks for future empirical validation
Will Compute Resources Be Provided?
Yes! Lambda is providing $400 in computing credits to each participating team, giving you access to powerful cloud instances (including A100s) to bring your ideas to life.
What Challenge Track Should I Choose?
You must align your project with one of our two tracks:
Project Track: This is a typical hackathon submission based on the starter materials or an original idea. We expect most of these to provide incremental theoretical progress and/or empirical evidence for (or against) an existing idea.
Relevant for higher context participants (i.e. those with experience running experiments on NNs or with an existing physics-AI agenda they want to accelerate progress on).
Exploration Track: Part of the idea for this hackathon is to open up and test the limits of a ‘physics for AI safety’ opportunity space. We have thought of some, but likely many more that would be exciting. Instead of traditional ‘hacking’, this will be a written submission, for example:
A proposal, outlining a novel physics-based solution to a key AI safety problem, with supporting material drawn from prior work.
A distillation or literature review bridging a CS/ML/AI idea with one from physics, with clear reference to an AI safety problem area.
We think this track has the potential to brainstorm some truly innovative ideas, while also allowing lower-context participants to get involved. For example, a ‘papers’ participant or group could partner with a ‘project’ group, resulting in a small research gain and a better understanding of the bigger picture.
What Problem Area Should I Focus On?
Your project should address one or more of our five problem areas:
Problem Area 1: Theory/Practice Gap for AI Interpretability - Apply physics methodologies (renormalization, computational mechanics, statistical mechanics) to build theoretical foundations for mechanistic interpretability
Problem Area 2: Learning, Inductive Biases, and Generalization - We want to understand how architecture, initialization, data distribution, and training procedures give rise to inductive biases—and in turn, how these biases control what is learned, when generalization occurs, and why models may fail.
Problem Area 3: Mathematical Models of Data Structure - AI alignment faces the challenge of understanding an AI's internal world model to prevent harmful actions. Physics methods, like coarse-graining and renormalization, could provide mathematical models to describe AI's learned representations and improve our understanding of AI's abstraction, generalization, and world modeling.
Problem Area 4: Quantitative Bounds on AI Behavior - This challenge focuses on using physics-inspired uncertainty quantification to establish quantitative bounds on AI behavior, ensuring alignment even when systems develop unanticipated strategies. Your work is crucial for guaranteeing ethical and reliable AI.
Problem Area 5: Scaling Laws, Phase Transitions, and Emergence - Uncover the physical foundations of AI performance, universal limits, and how smooth scaling reconciles with emergent capabilities.
Sign Ups
Entries
Overview
Resources
Guidelines
Schedule
Entries
Overview

This Challenge has been concluded
Over the July 25th-27th weekend intensive, participants tackled critical challenges by applying physics methodologies—from thermodynamics and statistical mechanics to computational mechanics—to understand and control AI systems. The event featured both Project and Exploration tracks, encouraging participants to bridge theoretical physics insights with practical AI safety solutions through collaborative innovation.
Following are the prize winners:
First Prize: Constrained Belief Updates in Transformer Representations:
Brianna Grado-White
This work extended computational mechanics theory to analyze how transformers implement optimal prediction under constraints. It showed belief state geometry across layers and that attention patterns decay exponentially with distance. Despite complex challenges, it bridged physics-inspired theory and practical interpretability, helping understand AI reasoning.
Second Prize: Momentum-Point-Perplexity Mechanics in Large Language Models
by Lorenzo Tomaz, Thomas Jones
This work found that energy remains approximately conserved across 20 different models. Their theoretical insights led to Jacobian steering—a control method that improves output quality while respecting the underlying physics, offering both fundamental understanding and practical applications for AI safety.Third Prize: Local Learning Coefficients in Reinforcement Learning:
by Jeremias Ferrao, Ilija LichkovskiThis project applies Singular Learning Theory to reinforcement learning, demonstrating how Local Learning Coefficients (LLCs) detect training phase transitions. By tracking reward components directly rather than derived metrics, LLC spikes correlate with capability milestones. This scalable approach enhances the understanding of how AI systems learn complex behaviors, with significant safety implications.
AI agentic system epidemiology
by Valentina Schastlivaia, Aray Karjauv
Using Physics-Informed Neural Networks to solve the governing equations, they created a framework for monitoring AI system health and predicting intervention effectiveness—providing a novel lens for understanding systemic AI risks.
Please find the link to the presentation here: https://docs.google.com/presentation/d/1IWKW9gMtc2qjfQAdWlRcT0EnJzeuJX3h6XOZC2oPOT4/edit?usp=sharing
Uniting Physics to Solve AI Safety
Join us for this AI Safety x Physics Grand Challenge to use your valuable skillset on chosen problem spaces for physicists!
You'll work alongside peers to create an original research project during this two day sprint with the guidance of expert mentors in the unique intersection of physics and AI safety. Be ready for excitement, progress, and, last but not least, prizes!
Physics and AI Safety?
The complexity of neural networks, the topology of their latent space, and the dynamics of the systems that deploy AGI today is an ideal topic of study for any physicist. We are at the absolute frontier of understanding learning dynamics through fundamental physics and during this event, you will become an integral part of this field's future!
Research Tracks

Employ your ML / AI experience from interpretability, benchmarking, or training to develop an original project employing the novel tools that physicists have introduced to the field in the past years.
Examples include the projects we've defined, such as operationalizing the natural abstraction hypothesis for empirical tests of compression or predictability of learned abstractions.

Propose novel research agendas to tackle the most pressing AI challenges or explore a potential bridge between an AI safety problem and a physics concept.
An example could be a proposal, outlining a novel physics-based solution to a key AI safety problem, with supporting material drawn from prior work.
Join
And experience…
Expert mentorship from leaders in the field of physics and AI safety
Structured problem areas with starter materials and concrete research directions
Global community of highly talented participants who are ambitiously solving top challenges
Follow-up support for exceptional projects through the Apart Research programs
Optional pre-event orientation and brainstorming for those new to AI safety
Shape the Future!
Join us in applying your academic training to one of the most important technical challenges of our time. Connect with a growing community of researchers safeguarding AI systems for the benefit of humanity.
FAQ
What Background Knowledge Is Required?
This challenge welcomes participants with physics backgrounds who are interested in applying rigorous methodologies to AI safety. While familiarity with machine learning is helpful, we encourage participation from:
Physics PhD students, postdocs, and faculty across all subfields
Theoretical physicists interested in complex systems and statistical mechanics
Experimental physicists with experience in data analysis and modeling
Computational physicists familiar with simulation and numerical methods
Physicists working at the intersection of physics and computer science
Anyone passionate about applying physics thinking to beneficial AI development
How Are Teams Formed?
Teams will self-organize during the first day through structured activities and interest-matching sessions. You can also form teams in advance with other registered participants via our Discord channel: https://ais.pub/disc
Teams should ideally combine:
Different physics backgrounds (theory, experiment, computation)
Varied experience levels with AI/ML
Complementary skills in mathematics, programming, and research
What Are The Expected Outputs?
Teams should produce research outputs and reports that demonstrate progress toward applying physics methodologies to AI safety challenges. The most promising projects may evolve into:
Academic papers or technical reports
Continued research collaboration with physics-AI safety researchers
Integration into ongoing research programs at leading institutions
Follow-up funding opportunities for extended research
Theoretical frameworks for future empirical validation
Will Compute Resources Be Provided?
Yes! Lambda is providing $400 in computing credits to each participating team, giving you access to powerful cloud instances (including A100s) to bring your ideas to life.
What Challenge Track Should I Choose?
You must align your project with one of our two tracks:
Project Track: This is a typical hackathon submission based on the starter materials or an original idea. We expect most of these to provide incremental theoretical progress and/or empirical evidence for (or against) an existing idea.
Relevant for higher context participants (i.e. those with experience running experiments on NNs or with an existing physics-AI agenda they want to accelerate progress on).
Exploration Track: Part of the idea for this hackathon is to open up and test the limits of a ‘physics for AI safety’ opportunity space. We have thought of some, but likely many more that would be exciting. Instead of traditional ‘hacking’, this will be a written submission, for example:
A proposal, outlining a novel physics-based solution to a key AI safety problem, with supporting material drawn from prior work.
A distillation or literature review bridging a CS/ML/AI idea with one from physics, with clear reference to an AI safety problem area.
We think this track has the potential to brainstorm some truly innovative ideas, while also allowing lower-context participants to get involved. For example, a ‘papers’ participant or group could partner with a ‘project’ group, resulting in a small research gain and a better understanding of the bigger picture.
What Problem Area Should I Focus On?
Your project should address one or more of our five problem areas:
Problem Area 1: Theory/Practice Gap for AI Interpretability - Apply physics methodologies (renormalization, computational mechanics, statistical mechanics) to build theoretical foundations for mechanistic interpretability
Problem Area 2: Learning, Inductive Biases, and Generalization - We want to understand how architecture, initialization, data distribution, and training procedures give rise to inductive biases—and in turn, how these biases control what is learned, when generalization occurs, and why models may fail.
Problem Area 3: Mathematical Models of Data Structure - AI alignment faces the challenge of understanding an AI's internal world model to prevent harmful actions. Physics methods, like coarse-graining and renormalization, could provide mathematical models to describe AI's learned representations and improve our understanding of AI's abstraction, generalization, and world modeling.
Problem Area 4: Quantitative Bounds on AI Behavior - This challenge focuses on using physics-inspired uncertainty quantification to establish quantitative bounds on AI behavior, ensuring alignment even when systems develop unanticipated strategies. Your work is crucial for guaranteeing ethical and reliable AI.
Problem Area 5: Scaling Laws, Phase Transitions, and Emergence - Uncover the physical foundations of AI performance, universal limits, and how smooth scaling reconciles with emergent capabilities.
Speakers & Collaborators
Jason Hoelscher-Obermaier
Organizer & Judge
Jason is co-director of Apart Research and leads Apart Lab, the research program supporting top hackathon participants and projects.
Ari Brill
Area Chair
Astrophysicist turned AI safety researcher. PhD in Physics from Columbia, former NASA postdoc studying black holes. Now develops mathematical models for AI system representations and alignment.
Lauren Greenspan
Area Chair
An interdisciplinary researcher bridging theoretical physics, social studies of science, and AI safety, Lauren works on the PIBBSS horizon scanning team. She is currently focused on research and field building to close the theory-practice gap in AI safety.
Paul Riechers
Speaker
Theoretical physicist and Co-founder of BITS and Simplex AI Safety. Research Lead at Astera Institute, focusing on the intersection of physics theory and AI alignment research.
Dmitry Vaintrob
Speaker
Dmitry Vaintrob is an AI safety and interpretability researcher. He has a mathematics background, having studied interactions between algebraic geometry, representation theory and physics. He is now working at the Horizon Scanning Team at PIBBSS
Jesse H
Speaker
Executive Director of Timaeus, leading research on singular learning theory and developmental interpretability for AI safety. Former research assistant at University of Cambridge studying the science of deep learning.
Daniel Kunin
Mentor
PhD candidate at the Institute for Computational and Mathematical Engineering at Stanford University, advised by Surya Ganguli.
Andrew Mack
Mentor
Andrew is an independent AI safety / interpretability researcher. Formerly senior engineer and postdoctoral fellow at Stanford with a PhD from Princeton.
Eric Michaud
Mentor
4th year PhD student in the Department of Physics at MIT and, as of June 2025, a resident at Goodfire AI.
Adam scherlis
Mentor
AI safety researcher (since 2020), with a particular interest in interpretability and related topics. Previously, Adam was a theoretical physicist, primarily focused on building and testing models of dark matter, which is composed of a vast number of extremely lightweight particles (e.g., axions).
Garrett Merz
Mentor
Garrett’s current work at DSI focuses on the development of AI for the computation of high-energy physics scattering amplitudes. Prior to coming to DSI, Garrett did his PhD on the ATLAS experiment at CERN, where he played a major role in the first direct measurement of the top-quark Higgs coupling
Martin Biehl
Mentor
Martin Biehl is senior research scientist at Cross Labs. His main research interest are the foundations of agency and the relation between physics and biology. He employs tools from probability theory, dynamical systems theory, and decision theory.
Esben Kran
Organizer
Esben is the CEO and Chariman of Apart. He has published award-winning AI safety research in various domains related to cybersecurity, autonomy preservation, and interpretability. He is involved in numerous efforts to ensure AI remains safe for humanity.
Archana Vaidheeswaran
Organizer
Archana is community manager at Apart. She is a board member at Women in Machine Learning and is a contributor to the TinyML community.
Speakers & Collaborators

Jason Hoelscher-Obermaier
Organizer & Judge
Jason is co-director of Apart Research and leads Apart Lab, the research program supporting top hackathon participants and projects.

Ari Brill
Area Chair
Astrophysicist turned AI safety researcher. PhD in Physics from Columbia, former NASA postdoc studying black holes. Now develops mathematical models for AI system representations and alignment.

Lauren Greenspan
Area Chair
An interdisciplinary researcher bridging theoretical physics, social studies of science, and AI safety, Lauren works on the PIBBSS horizon scanning team. She is currently focused on research and field building to close the theory-practice gap in AI safety.

Paul Riechers
Speaker
Theoretical physicist and Co-founder of BITS and Simplex AI Safety. Research Lead at Astera Institute, focusing on the intersection of physics theory and AI alignment research.

Dmitry Vaintrob
Speaker
Dmitry Vaintrob is an AI safety and interpretability researcher. He has a mathematics background, having studied interactions between algebraic geometry, representation theory and physics. He is now working at the Horizon Scanning Team at PIBBSS

Jesse H
Speaker
Executive Director of Timaeus, leading research on singular learning theory and developmental interpretability for AI safety. Former research assistant at University of Cambridge studying the science of deep learning.

Daniel Kunin
Mentor
PhD candidate at the Institute for Computational and Mathematical Engineering at Stanford University, advised by Surya Ganguli.

Andrew Mack
Mentor
Andrew is an independent AI safety / interpretability researcher. Formerly senior engineer and postdoctoral fellow at Stanford with a PhD from Princeton.

Eric Michaud
Mentor
4th year PhD student in the Department of Physics at MIT and, as of June 2025, a resident at Goodfire AI.

Adam scherlis
Mentor
AI safety researcher (since 2020), with a particular interest in interpretability and related topics. Previously, Adam was a theoretical physicist, primarily focused on building and testing models of dark matter, which is composed of a vast number of extremely lightweight particles (e.g., axions).

Garrett Merz
Mentor
Garrett’s current work at DSI focuses on the development of AI for the computation of high-energy physics scattering amplitudes. Prior to coming to DSI, Garrett did his PhD on the ATLAS experiment at CERN, where he played a major role in the first direct measurement of the top-quark Higgs coupling

Martin Biehl
Mentor
Martin Biehl is senior research scientist at Cross Labs. His main research interest are the foundations of agency and the relation between physics and biology. He employs tools from probability theory, dynamical systems theory, and decision theory.

Esben Kran
Organizer
Esben is the CEO and Chariman of Apart. He has published award-winning AI safety research in various domains related to cybersecurity, autonomy preservation, and interpretability. He is involved in numerous efforts to ensure AI remains safe for humanity.
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Registered Jam Sites
Register A Location
Beside the remote and virtual participation, our amazing organizers also host local hackathon locations where you can meet up in-person and connect with others in your area.
The in-person events for the Apart Sprints are run by passionate individuals just like you! We organize the schedule, speakers, and starter templates, and you can focus on engaging your local research, student, and engineering community.
Our Other Sprints
Jun 13, 2025
-
Jun 13, 2025
Research
Red Teaming A Narrow Path: ControlAI Policy Sprint
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up
May 30, 2025
-
Jun 1, 2025
Research
Apart x Martian Mechanistic Router Interpretability Hackathon
This unique event brings together diverse perspectives to tackle crucial challenges in AI alignment, governance, and safety. Work alongside leading experts, develop innovative solutions, and help shape the future of responsible
Sign Up
Sign Up
Sign Up

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events