This work was done during one weekend by research workshop participants and does not represent the work of Apart Research.
ApartSprints
Howard University AI Safety Summit & Policy Hackathon
67240c8fb0416a4520d2b4b6
Howard University AI Safety Summit & Policy Hackathon
November 21, 2024
Accepted at the 
67240c8fb0416a4520d2b4b6
 research sprint on 

User Transparency Within AI

Generative AI technologies present immense opportunities but also pose significant challenges, particularly in combating misinformation and ensuring ethical use. This policy paper introduces a dual-output transparency framework requiring organizations to disclose AI-generated content clearly. The proposed system provides users with a choice between fact-based and mixed outputs, both accompanied by clear markers for AI generation. This approach ensures informed user interactions, fosters intellectual integrity, and aligns AI innovation with societal trust. By combining policy mandates with technical implementations, the framework addresses the challenges of misinformation and accountability in generative AI.

By 
Jonathan King, Robert Hardy, Jeremiah Bailey, Amir Abdulgadir
🏆 
4th place
3rd place
2nd place
1st place
 by peer review
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

This project is private