Sep 15, 2025
Towards Agnostic Viral Engineering Detection
Aaron Maiwald, Hanna Palya, Alejandro Acelas
Genetic engineering tools could potentially be misused to createharmful pathogens, making early detection of engineered virusescritical for biosecurity. We developed an agnostic geneticengineering detection system for viruses, simulating three types ofmodifications (deletions, inversions, and frameshift mutations)across 25 human-infecting viral genomes to create acomprehensive synthetic dataset. We benchmarked threeclassification approaches—k-mer-based logistic regression,BLAST alignment, and convolutional neural networks—alongsidean ensemble method. All models achieved F1-scores below 7%,suggesting that standard bioinformatics and machine learningapproaches are insufficient for robust detection of diverse viralengineering signatures.
This was a strong project. For the time allowed in this sprint- it's extremely well thought out and functional. There have been areas where biological oversimplifications have been implemented, it may not be ready to inform real threat detection but is an effective baseline probe. Reproducibility and documentation needs to be improved, but limitations and results have been clearly highlighted.
I would also recommend mapping this project to broader AI safety mechanisms to bridge the policy-practice gap that currently exists in the CBRN space.
This project is highly impressive. It clearly addresses a critical gap in biodefense capabilities, detecting engineered viruses. The team built a novel synthetic dataset that includes engineering types not previously studied for early detection (such as deletions and inversions), tested several approaches, and convincingly showed that standard methods don’t work well, which is a valuable finding. The write-up is clear, well-documented, and reproducible. While it doesn’t yet provide a scalable new detection tool, it highlights a crucial research gap and lays strong groundwork for future progress.
This is an ambitious extension of NAO’s work, with clear background explanations and providing useful negative results suggesting that simple baseline methods don’t succeed, which is always important to check first. The inclusion of two different train–test splits is a good idea and shows care for potential biases, though the results for the second method aren’t shown.
At the same time, results like BLAST performing worse than random suggest that something may be off and needs to be checked carefully. The conclusion that detection is inherently hard feels premature, since NAO obtained good F1 scores on chimeras that you did not replicate, even with similar datasets and methods.
A good next step would be to carefully debug and validate each part of the pipeline before moving on to more complex methods, and to improve the documentation so the work is easier to follow. Reaching out to NAO’s team to cross-check results could also be helpful. Once the basics are more solid, the project could explore advanced methods (for example genomic foundation models?).
Regarding code replication, I quickly checked the code and managed to rerun the notebooks.
Cite this work
@misc {
title={
(HckPrj) Towards Agnostic Viral Engineering Detection
},
author={
Aaron Maiwald, Hanna Palya, Alejandro Acelas
},
date={
9/15/25
},
organization={Apart Research},
note={Research submission to the research sprint hosted by Apart.},
howpublished={https://apartresearch.com}
}


