May 6, 2024
LEGISLaiTOR: A tool for jailbreaking the legislative process
Willie Chalmers III, Margaret Belford
In this work, we consider the ramifications on generative artificial intelligence (AI) tools in the legislative process in democratic governments. While other research focuses on the micro-level details associated with specific models, this project takes a macro-level approach to understanding how AI can assist the legislative process.
Esben Kran
Love the concept of Trojan Microlegislation and an obvious path for malicious use of AI or rogue AGI to inject long-term changes to society without politicians knowing this with a great example in Robert Moses. The example tool unfortunately wasn't live during the review process but it looks like a great demonstration to uncover some of these risks in-action. For continuing the work, a few ideas could be to integrate this into how prompt injection works in current technical systems, how more previous examples have looked, and developing the legislative defenses for this in a context that can be used across the US. From an AI safety perspective, this is also where it's relevant to figure out which sorts of laws we want to look out for, somewhat similar to spot-checking codebases for suspicious API requests if an AI has created it (esbenkc/karnak) or something similar.
Bart Bussmann
Cool project! Unfortunately, the demo doesn’t seem to be working at this moment. That’s a shame, because based on the description it seems really cool.
I think the trojan microregulation is an interesting and realistic treat scenario that is very underexplored at the moment. If you can get an LLM to do this well, I think this project has great potential to be turned into a paper!
Andrey Anurin
Impressive and engaging writeup! Investigating potential uses of AI in the legislation creation process seems very timely, I imagine there’s already hundreds of low-level staffers injecting delve-ridden texts into bills all over the US. Both the process and the democratic system itself will have to adjust to mitigate the risks. A direction in which this work could be extended is trying to see how real human beings interact with {malicious-,benign-}{AI,human} produced legislation.
Sadly, I could access neither the deployed app nor the sources.
Cite this work
@misc {
title={
@misc {
},
author={
Willie Chalmers III, Margaret Belford
},
date={
5/6/24
},
organization={Apart Research},
note={Research submission to the research sprint hosted by Apart.},
howpublished={https://apartresearch.com}
}