This work was done during one weekend by research workshop participants and does not represent the work of Apart Research.
ApartSprints
Deception Detection Hackathon: Preventing AI deception
660d65646a619f5cf53b1f56
Deception Detection Hackathon: Preventing AI deception
July 1, 2024
Accepted at the 
660d65646a619f5cf53b1f56
 research sprint on 

Evaluating and inducing steganography in LLMs

This report demonstrates that large language models are capable of hiding simple 8 bit information in their output using associations from more powerful overseers (other LLMs or humans). Without direct steganography fine tuning, LLAMA 3 8B can guess a 8 bit hidden message in a plain text in most cases (69%), however a more capable model, GPT-3.5 was able to catch almost all of them (84%). More research is required to investigate how this ability might be induced or improved via RL training in similar and larger models.

By 
Artem Karpov
🏆 
4th place
3rd place
2nd place
1st place
 by peer review
Thank you! Your submission is under review.
Oops! Something went wrong while submitting the form.

This project is private