Jul 29, 2024
Data Massager
Giles Edkins, Ayush Jain
A VSCode plugin for helping the creation of Q&A datasets used for the evaluation of LLMs capabilities and alignment.
Marc Carauleanu
Great project! I really like the wide array of features that are well thought out from the perspective of the needs of the user and the integration with VS Code! I would be excited to see features that help automate the data generation and curation pipeline if the steps are similar to what the user has done in the past.
Jonny Spicer
I like this project a lot, and I can see it really shining when using large models to create datasets for novel benchmarks for smaller models. I like how intuitive the UI in VSCode is, and that the prospective dataset creator doesn’t have to leave their IDE to use the tool. As a further step, perhaps you could consider prompting the LLM to give a confidence score for its answer - humans could then review all answers below a certain confidence score, or simply choose to remove them all from the dataset.
Jason Schreiber
This project seems very promising and the writeup is very helpful for understanding the current scope, strengths and weaknesses. For me the most important question to address here is how to measure the reliability of the different operations on the fly and how to improve it. I think it will be very hard (and for some datasets maybe impossible in the near-term) to build a faultless system, so the core question is how to get a tool that is aware of its own limitations and lets the human researchers work around those limitations.
Jacques Thibodeau
I think the general idea for this project is good. There probably should have been more focus on figuring out how to make it uniquely good at coming up with new alignment evaluations (is the model being deceptive? is it sandbagging? does it have situational awareness?) and trying to make it especially good at helping the researcher coming it with new examples. Overall, I like the different features and think there is some shape of this kind of project that could make it significantly easier to generate higher quality evals.
Cite this work
@misc {
title={
@misc {
},
author={
Giles Edkins, Ayush Jain
},
date={
7/29/24
},
organization={Apart Research},
note={Research submission to the research sprint hosted by Apart.},
howpublished={https://apartresearch.com}
}