Apr 14, 2023

-

Apr 17, 2023

Online & In-Person

The Interpretability Hackathon 2.0

We focus on interpreting the innards of neural networks using methods developed from mechanistic interpretability

00:00:00:00

00:00:00:00

00:00:00:00

00:00:00:00

We focus on interpreting the innards of neural networks using methods developed from mechanistic interpretability

This event is ongoing.

This event has concluded.

32

Sign Ups

Overview

Resources

Schedule

Overview

Arrow

Hosted by Zaki, Apart Research, Esben Kran, StefanHex, calcan, Neel Nanda · #alignmentjam

Join this AI safety hackathon to find new perspectives on the "brains" of AI!

48 hours intense research in interpretability, the modern AI neuroscience.

We focus on interpreting the innards of neural networks using methods developed from mechanistic interpretability! With the starter Colab code, you should be able to quickly get a technical insight into the process.

We provide you with the best starter templates that you can work from so you can focus on creating interesting research instead of browsing Stack Overflow. You're also very welcome to check out some of the ideas already posted!

Join the hackathon Discord

How to participate

Create a user on the itch.io (this) website and click participate. We will assume that you are going to participate and ask you to please cancel if you won't be part of the hackathon.

Instructions

You will work on research ideas you generate during the hackathon and you can find more inspiration below.

Submission

Everyone will help rate the submissions together on a set of criteria that we ask everyone to follow. This will happen over a 2-week period and will be open to everyone!

32

Sign Ups

Overview

Resources

Schedule

Overview

Arrow

Hosted by Zaki, Apart Research, Esben Kran, StefanHex, calcan, Neel Nanda · #alignmentjam

Join this AI safety hackathon to find new perspectives on the "brains" of AI!

48 hours intense research in interpretability, the modern AI neuroscience.

We focus on interpreting the innards of neural networks using methods developed from mechanistic interpretability! With the starter Colab code, you should be able to quickly get a technical insight into the process.

We provide you with the best starter templates that you can work from so you can focus on creating interesting research instead of browsing Stack Overflow. You're also very welcome to check out some of the ideas already posted!

Join the hackathon Discord

How to participate

Create a user on the itch.io (this) website and click participate. We will assume that you are going to participate and ask you to please cancel if you won't be part of the hackathon.

Instructions

You will work on research ideas you generate during the hackathon and you can find more inspiration below.

Submission

Everyone will help rate the submissions together on a set of criteria that we ask everyone to follow. This will happen over a 2-week period and will be open to everyone!

32

Sign Ups

Overview

Resources

Schedule

Overview

Arrow

Hosted by Zaki, Apart Research, Esben Kran, StefanHex, calcan, Neel Nanda · #alignmentjam

Join this AI safety hackathon to find new perspectives on the "brains" of AI!

48 hours intense research in interpretability, the modern AI neuroscience.

We focus on interpreting the innards of neural networks using methods developed from mechanistic interpretability! With the starter Colab code, you should be able to quickly get a technical insight into the process.

We provide you with the best starter templates that you can work from so you can focus on creating interesting research instead of browsing Stack Overflow. You're also very welcome to check out some of the ideas already posted!

Join the hackathon Discord

How to participate

Create a user on the itch.io (this) website and click participate. We will assume that you are going to participate and ask you to please cancel if you won't be part of the hackathon.

Instructions

You will work on research ideas you generate during the hackathon and you can find more inspiration below.

Submission

Everyone will help rate the submissions together on a set of criteria that we ask everyone to follow. This will happen over a 2-week period and will be open to everyone!

32

Sign Ups

Overview

Resources

Schedule

Overview

Arrow

Hosted by Zaki, Apart Research, Esben Kran, StefanHex, calcan, Neel Nanda · #alignmentjam

Join this AI safety hackathon to find new perspectives on the "brains" of AI!

48 hours intense research in interpretability, the modern AI neuroscience.

We focus on interpreting the innards of neural networks using methods developed from mechanistic interpretability! With the starter Colab code, you should be able to quickly get a technical insight into the process.

We provide you with the best starter templates that you can work from so you can focus on creating interesting research instead of browsing Stack Overflow. You're also very welcome to check out some of the ideas already posted!

Join the hackathon Discord

How to participate

Create a user on the itch.io (this) website and click participate. We will assume that you are going to participate and ask you to please cancel if you won't be part of the hackathon.

Instructions

You will work on research ideas you generate during the hackathon and you can find more inspiration below.

Submission

Everyone will help rate the submissions together on a set of criteria that we ask everyone to follow. This will happen over a 2-week period and will be open to everyone!

Speakers & Collaborators

Neel Nanda

Speaker & Judge

Team lead for the mechanistic interpretability team at Google Deepmind and a prolific advocate for open source interpretability research.

Esben Kran

Organizer and Keynote Speaker

Esben is the co-director of Apart Research and specializes in organizing research teams on pivotal AI security questions.

Speakers & Collaborators

Neel Nanda

Speaker & Judge

Team lead for the mechanistic interpretability team at Google Deepmind and a prolific advocate for open source interpretability research.

Esben Kran

Organizer and Keynote Speaker

Esben is the co-director of Apart Research and specializes in organizing research teams on pivotal AI security questions.