Feb 1, 2026
LINE⁴: Operationalizing the Edge of AI Risk
Adjoa Sarpong, Joseph Awuah, Amo Joshua Selorm
LINE⁴ (Line Four) aggregates official safety evaluations across three frameworks to provide transparent, real-time visibility into AI safety assessments across major labs. The dashboard tracks risks across four critical dimensions: CBRN proliferation, cyber offense capabilities, autonomous replication, and deceptive alignment.
Data Source: Real assessments extracted from official lab System Cards
Anthropic: Responsible Scaling Policy (RSP) with Automation Safety Levels (ASL)
OpenAI: Preparedness Framework with severity assessments
DeepMind: Frontier Safety Framework with Critical Capability Levels (CCL)
its true that the basic vibe of finding the exchange rate between different labs' disclosure vocabulary of capability levels is important and helpful!
Don't tell me you used streamlit and nextjs in the writeup. That's not a "method", and it spent valuable wordcount that could've been used to elaborate on how you drafted the exchange rates and LLM-as-judge'd over them (sections 3.1 and 3.2 are in my view the most critical, the parts of the paper that actually matter, and you were way too brief about them. I would've happily read more detail about how gemini auditor was prompted, for instance, plus discussion in 3.1 about how you derived buckets and reasons your bucketing might be wrong.
The team addresses a real problem with a reasonable idea but the work can only be used as a middle man between internal big labs' evaluations and public risk assessments. I can still imagine it being useful however.
Cite this work
@misc {
title={
(HckPrj) LINE⁴: Operationalizing the Edge of AI Risk
},
author={
Adjoa Sarpong, Joseph Awuah, Amo Joshua Selorm
},
date={
2/1/26
},
organization={Apart Research},
note={Research submission to the research sprint hosted by Apart.},
howpublished={https://apartresearch.com}
}


