Nov 22, 2024

-

Nov 25, 2024

Online & In-Person

Reprogramming AI Models Hackathon

Whether you're an AI researcher, a curious developer, or passionate about making AI systems more transparent and controllable, this hackathon is for you. As a participant, you will: Collaborate with experts to create novel AI observability tools Learn about mechanistic interpretability from industry leaders Contribute to solving real-world challenges in AI safety and reliability Compete for prizes and the opportunity to influence the future of AI development Register now and be part of the movement towards more transparent, reliable, and beneficial AI systems. We provide access to Goodfire's SDK/API and research preview playground, enabling participation regardless of prior experience with AI observability.

00:00:00:00

00:00:00:00

00:00:00:00

00:00:00:00

This event is ongoing.

This event has concluded.

Overview

Overview

Arrow
Arrow
Arrow

Why This Matters

As AI models become more powerful and widespread, understanding their internal mechanisms isn't just academic curiosity—it's crucial for building reliable, controllable AI systems. Mechanistic interpretability gives us the tools to peek inside these "black boxes" and understand how they actually work, neuron by neuron and feature by feature.

What You'll Get

  • Exclusive Access: Use Goodfire's API to access an interpretable 8B or 70B model with efficient inference.

  • Cutting-Edge Tools: Experience Goodfire's SDK/API for feature steering and manipulation

  • Advanced Capabilities: Work with conditional feature interventions and sophisticated development flows

  • Free Resources: Compute credits for every team to ensure you can pursue ambitious projects

  • Expert Guidance: Direct mentorship from industry leaders throughout the weekend

Project Tracks

1. Feature Investigation

  • Map and analyze feature phenomenology in large language models

  • Discover and validate useful feature interventions

  • Research the relationship between feature weights and intervention success

  • Develop metrics for intervention quality assessment

2. Tooling Development

  • Build tools for automated feature discovery

  • Create testing frameworks for intervention reliability

  • Develop integration tools for existing ML frameworks

  • Improve auto-interpretation techniques

3. Visualization & Interface

  • Design intuitive visualizations for feature maps

  • Create interactive tools for exploring model internals

  • Build dashboards for monitoring intervention effects

  • Develop user interfaces for feature manipulation

4. Novel Research

  • Investigate improvements to auto-interpretation

  • Study feature interaction patterns

  • Research intervention transfer between models

  • Explore new approaches to model steering

Why Goodfire's Tools?

While participants are welcome to use their existing setups, Goodfire's API brings exceptional value to this hackathon as a primary option for participants.

Goodfire provides:

  • Access to a 70B parameter model via API (with efficient inference)

  • Feature steering capabilities made simple through the SDK/API

  • Advanced development workflows including conditional feature interventions

The hackathon serves as a unique opportunity for Goodfire to gather valuable feedback from the developer community on their API/SDK. To ensure all participants can pursue ambitious research projects without constraints, Goodfire is providing free compute credits to every team.

Previous Participant Experiences

"I learned so much about AI Safety and Computational Mechanics. It is a field I have never heard of, and it combines two of my interests - AI and Physics. Through the hackathons, I gained valuable connections and learned a lot from researchers with extensive experience." - Doroteya Stoyanova, Computer Vision Intern

Resources

Resources

Arrow
Arrow
Arrow

To ensure you're well-equipped for the Reprogramming AI Models Hackathon, we've compiled a set of resources to support your participation:

  1. Goodfire's SDK/API with hosted inference: Your primary toolkit for the hackathon. Familiarize yourself with our framework for understanding and modifying AI model behavior.

    • Hosted inference on Llama 3 8B and 70B models

    • Feature inspection and intervention capabilities: https://docs.goodfire.ai/examples/advanced.html#Feature-intervention-modes

    • Example notebooks and tutorials: https://docs.goodfire.ai/examples/quickstart.html#Use-contrastive-features-to-fine-tune-with-a-single-example!

    • Latent explorer visualization tools: https://docs.goodfire.ai/examples/latent_explorer.html

    • Rate limits and usage guidelines: https://docs.goodfire.ai/rate-limits.html

  2. Research Preview Playground

    • Sandbox environment for model experimentation: https://docs.goodfire.ai/examples/quickstart.html#Replace-model-calls-with-OpenAI-compatible-API

    • Feature activation analysis tools: https://docs.goodfire.ai/examples/advanced.html#Feature-intervention-modes

    • Conditional intervention testing: https://docs.goodfire.ai/examples/advanced.html#Conditional-feature-interventions

  3. Check out the Jupyter Notebook Quickstart: . In this quickstart, you'll learn how to:

    • Sample from a language model (in this case, Llama 3 8B)

    • Search for exciting features and intervene in them to steer the model

    • Find features by contrastive search

    • Save and load Llama models with steering applied

  4. Tutorial: Visualizing AI Model Internals: Watch this video to understand how to use Goodfire's tools to map and visualize AI model behavior.

  1. The Cognitive Revolution Podcast - Episode on Interpretability. n this episode of The Cognitive Revolution, we delve into the science of understanding AI models' inner workings, recent breakthroughs, and the potential impact on AI safety and control

  2. Auto-interp Paper: This paper applies automation to the problem of scaling an interpretability technique to all the neurons in a large language model.

  3. ARENA Interpretability with SAEs

  4. Gemma Scope: a comprehensive, open suite of sparse autoencoders for language model interpretability.

  5. Neuronpedia: Platform for accelerating research into Sparse Autoencoders

  6. The Geometry of Concepts: Sparse Autoencoder Feature Structure Paper. This paper investigates the structured organization of concept representations within large language models using sparse autoencoders, revealing a multi-scale structure with refined atomic parallelogram forms, modular brain-like spatial features, and anisotropic galaxy-scale distributions with unique eigenvalue properties.

  7. Open Source Replication of Anthropic’s Crosscoder paper for model-diffing

  8. Lesswrong search for SAE

Schedule

Schedule

Arrow
Arrow
Arrow

Here is the schedule for the Hackathon:
We start with an introductory talk and end the event during the following week with an awards ceremony. Join the public ICal here. You will also find Explorer events, such as collaborative brainstorming and team match-making before the hackathon begins on Discord and in the calendar.

Entries

Can we steer a model’s behavior with just one prompt? investigating SAE-driven auto-steering

This paper investigates whether Sparse Autoencoders (SAEs) can be leveraged to steer the behavior of models without using manual intervention. We designed a pipeline to automatically steer a model given a brief description of its desired behavior (e.g.: “Behave like a dog”). The pipeline is as follows: 1. We automatically retrieve behavior-relevant SAE features. 2. We choose an input prompt (e.g.: “What would you do if I gave you a bone?” or “How are you?”) over which we evaluate the model’s responses. 3. Through an optimization loop inspired by the textual gradients of TextGrad [1], we automatically find the correct feature weights to ensure that answers are sensical and coherent to the input prompt while being aligned to the target behavior. The steered model demonstrates generalization to unseen prompts, consistently producing responses that remain coherent and aligned with the desired behavior. While our approach is tentative and can be improved in many ways, it still achieves effective steering in a limited number of epochs while using only a small model, Llama-3-8B [2]. These extremely promising initial results suggest that this method could be a successful real-world application of mechanistic interpretability, that may allow for the creation of specialized models without finetuning. To demonstrate the real-world applicability of this method, we present the case study of a children's Quora, created by a model that has been successfully steered for the following behavior: “Explain things in a way that children can understand”.

Learn More

Speakers & Collaborators

Tom McGrath

Organizer & Judge

Chief Scientist at Goodfire, previously Senior Research Scientist at Google DeepMind, where he co-founded the interpretability team

Neel Nanda

Speaker & Judge

Team lead for the mechanistic interpretability team at Google Deepmind and a prolific advocate for open source interpretability research.

Dan Balsam

Organizer & Mentor

CTO at Goodfire, previously Founding Engineer and Head of AI at RippleMatch. Goodfire makes Interpretability products for safe and reliable generative AI models.

Myra Deng

Organizer

Founding PM at Goodfire, Stanford MBA and MS CS graduate previously building modeling platforms at Two Sigma

Archana Vaidheeswaran

Organizer

Archana is responsible for organizing the Apart Sprints, research hackathons to solve the most important questions in AI safety.

Jaime Raldua

Organiser

Jaime has 8+ years of experience in the tech industry. Started his own data consultancy to support EA Organisations and currently works at Apart Research as Research Engineer.

Joseph Bloom

Speaker

Joseph co-founded Decode Research, a non-profit organization aiming to accelerate progress in AI safety research infrastructure, and is a mechanistic interpretability researcher.

Alana X

Judge

Member of Technical Staff at Magic, leading initiatives on model evaluations. Previously a Research Intern at METR

Callum McDougall

Speaker

ARENA Director and SERI-MATS alumnus specializing in mechanistic interpretability and AI alignment education

Mateusz Dziemian

Judge

Member of Technical Staff at Gray Swan AI, recently worked on a U.K AISI collaboration. Previous participant in an Apart Sprint which ended in a NeurIPS workshop paper

Simon Lermen

Judge

MATS Scholar under Jeffrey Ladish and is an independent researcher. Mentored a spar project on AI agents and worked on spear-phishing people with AI agents.

Liv Gorton

Judge

Founding Research Scientist at Goodfire AI. Undertook independent research on sparse autoencoders in InceptionV1

Esben Kran

Organizer

Esben is the co-director of Apart Research and specializes in organizing research teams on pivotal AI security questions.

Jason Schreiber

Organizer and Judge

Jason is co-director of Apart Research and leads Apart Lab, our remote-first AI safety research fellowship.

Speakers & Collaborators

Tom McGrath

Organizer & Judge

Chief Scientist at Goodfire, previously Senior Research Scientist at Google DeepMind, where he co-founded the interpretability team

Neel Nanda

Speaker & Judge

Team lead for the mechanistic interpretability team at Google Deepmind and a prolific advocate for open source interpretability research.

Dan Balsam

Organizer & Mentor

CTO at Goodfire, previously Founding Engineer and Head of AI at RippleMatch. Goodfire makes Interpretability products for safe and reliable generative AI models.

Myra Deng

Organizer

Founding PM at Goodfire, Stanford MBA and MS CS graduate previously building modeling platforms at Two Sigma

Archana Vaidheeswaran

Organizer

Archana is responsible for organizing the Apart Sprints, research hackathons to solve the most important questions in AI safety.

Jaime Raldua

Organiser

Jaime has 8+ years of experience in the tech industry. Started his own data consultancy to support EA Organisations and currently works at Apart Research as Research Engineer.

Joseph Bloom

Speaker

Joseph co-founded Decode Research, a non-profit organization aiming to accelerate progress in AI safety research infrastructure, and is a mechanistic interpretability researcher.

Alana X

Judge

Member of Technical Staff at Magic, leading initiatives on model evaluations. Previously a Research Intern at METR

Callum McDougall

Speaker

ARENA Director and SERI-MATS alumnus specializing in mechanistic interpretability and AI alignment education

Mateusz Dziemian

Judge

Member of Technical Staff at Gray Swan AI, recently worked on a U.K AISI collaboration. Previous participant in an Apart Sprint which ended in a NeurIPS workshop paper

Simon Lermen

Judge

MATS Scholar under Jeffrey Ladish and is an independent researcher. Mentored a spar project on AI agents and worked on spear-phishing people with AI agents.

Liv Gorton

Judge

Founding Research Scientist at Goodfire AI. Undertook independent research on sparse autoencoders in InceptionV1

Esben Kran

Organizer

Esben is the co-director of Apart Research and specializes in organizing research teams on pivotal AI security questions.

Jason Schreiber

Organizer and Judge

Jason is co-director of Apart Research and leads Apart Lab, our remote-first AI safety research fellowship.