Feb 2, 2026
AI Safety Threshold Tracker
Revathi Prasad
AI Safety Threshold Tracker
An interactive dashboard that connects AI capability benchmarks to governance-defined thresholds. The tracker aggregates verified data from third-party evaluations (METR, SWE-Bench, WMDP) and company RSP disclosures (Anthropic ASL classifications, CBRN uplift scores), mapping them to thresholds derived from international frameworks including IDAIS Red Lines, EU AI Act, and Seoul Summit commitments.
Key features:
1. User-configurable thresholds with transparent justification labeling (distinguishing data-driven thresholds from interpretations)
2. 4 active capability metrics, 2 categorical RSP metrics, 5 inactive metrics awaiting data
3. Real-time tracking of frontier model capabilities against policy-relevant thresholds
This is a practical project that fills a real gap: turning vague “threshold” talk into something you can actually track. The framing is sensible and the dashboard is clear.
The main limitation is that the threshold mappings are still somewhat heuristic. This is more “good governance plumbing” than a new safety method. The impact depends on adoption.
To level up, I’d want: backtesting (“would this have changed decisions historically?”), uncertainty / missing-data handling, and a clean process for updating thresholds without Goodharting.
A useful tool with a simple idea. Only a few models has been tested which is reasonable given the time constrains.
Cite this work
@misc {
title={
(HckPrj) AI Safety Threshold Tracker
},
author={
Revathi Prasad
},
date={
2/2/26
},
organization={Apart Research},
note={Research submission to the research sprint hosted by Apart.},
howpublished={https://apartresearch.com}
}


