Interesting and clear paper! A few comments:
- Could have tested with more than 5 models and assess the effectiveness of the pipeline on varying degrees of modification (e.g. benign fine-tunes that make large weight changes)
- Talk more in-depth about which phase of the classifier does the work (thresholds heuristics? the random forest? something else?). This seems to be the core of the contribution here and we don’t learn much about it
- Confidence scores could be improved; 52-68% seems little and would currently lead to many false positives or negatives.
Like the angle! Hf model security is likely going to be a huge problem. Thought the cheap static check into expensive dynamic check was a good touch to make it actually scalable. It would be great to see more models and a bit more details on the method but for a hackathon sprint its a nice proof-of-concept!
Cite this work
@misc {
title={
(HckPrj) AUDIT:
},
author={
Kevin Zhang, Derrick Yao, Yogesh Prabhu, Bryan Zhang
},
date={
2/1/26
},
organization={Apart Research},
note={Research submission to the research sprint hosted by Apart.},
howpublished={https://apartresearch.com}
}


