Shape the future of safe and ethical AI development! Whether you're a researcher, developer, policy enthusiast, or new to AI safety - join us for an empowering weekend of innovation and collaboration. No prior AI safety experience required. Together, we can build the foundational tools and frameworks needed for responsible AI development.
We're thrilled to partner with Women Who Do Data (W2D2), a member-led community dedicated to supporting diverse talent in AI technology development, to bring you this unique hackathon during International Women's Day weekend. Together, we're working to increase the presence of underrepresented groups in AI safety and technical AI development.
We're excited to announce that Lambda Labs will be providing $400 in computing credits to each participating team! This generous support will give you access to powerful cloud instances (including A100s) to help bring your ideas to life. Each track will have a $600 prize for the winning team, with a total of $1,800 in prizes across all three tracks!
The Women in AI Safety Hackathon brings together talented individuals to tackle crucial challenges in AI development and deployment. This event particularly encourages women and underrepresented groups to contribute their unique perspectives to critical areas of AI safety, including alignment, governance, security, and evaluation.
As AI systems become increasingly powerful and pervasive, diverse perspectives in their development and safety mechanisms are more crucial than ever. This hackathon provides a platform for participants to:
We're excited to announce that Goodfire, a pioneering research lab in AI interpretability, will be sponsoring our Mechanistic Interpretability track! We're also thrilled to have Myra Deng, Founding PM at Goodfire, join us for a HackTalk on understanding and steering AI models.
1. Understanding AI Model Internals: Dive deep into the inner workings of large language models using state-of-the-art sparse autoencoder techniques. Learn to map and understand model behavior at a granular level, while developing tools to interpret and visualize neural features. This track focuses on creating innovative approaches to understand what's happening inside these complex systems.
2. Model Steering and Editing: Explore practical applications of interpretability by developing methods to modify and control model behavior through targeted feature interventions. Work on creating interpretable control mechanisms and safety-focused editing tools that can help shape model outputs while maintaining performance. This component bridges the gap between theoretical understanding and practical applications.
3. Feature Analysis and Visualization: Create tools and interfaces that make AI systems more transparent and understandable. Focus on building interactive visualization dashboards that help analyze feature activations and interactions, making complex neural networks more accessible to researchers and practitioners. Develop novel ways to present and interact with model internals
Participants will have access to:
- Goodfire's Ember API for model interpretation
- Compute credits for experiments
The winning team in this track will receive a $600 prize!
Build an education platform feature to help newcomers to the field understand the risks of AI.
Ideas to get you started include:
The winning team in this track will receive a $600 prize!
This track invites participants to explore the intersection of Artificial Intelligence and the Social Sciences through two broad themes: (1) how AI "thinks" and (2) how it shapes human society.
The first theme, Machine Psychology, examines the systematic study of AI behaviour and cognition, drawing insights from psychology, behavioural science, and cognitive science. This theme investigates how AI models develop internal representations, exhibit emergent capabilities, display decision-making patterns, and adapt to different inputs. It explores both the interpretability of these systems and their behavioural tendencies, helping us understand how AI "thinks”.
The second theme, AI in Society, focuses on the human-AI dynamic, investigating AI’s broader social, political, and economic implications. This includes its impact on human interactions, institutional transformation across sectors like education and healthcare, and evolving social norms. The theme also addresses critical ethical considerations in AI development and deployment, examining questions of fairness, accountability, and governance as AI systems become more integrated into society.
Participants will contribute through short reports or research-based papers that can range from highly technical to completely non-technical, as we want to make this track accessible to all social scientists and foster interdisciplinary discussions on AI’s evolving role in our world.
The winning team in this track will receive a $600 prize!
️
1. Intro to Transformative AI curriculum: The risks and opportunities of advanced AI are evolving at unprecedented speed — and so is the need for capable individuals to shape its trajectory. This course is for those who want to rapidly develop their understanding of transformative AI and its impact on humanity. Through expert-facilitated discussions and carefully curated materials, you’ll explore the technical foundations of AI, examine potential futures, and debate key ideas alongside others passionate about ensuring AI benefits humanity. By the end, you’ll have both the knowledge and network to take meaningful steps toward contributing to AI safety
2. BlueDot's science of learning by Li-Lian Ang: This article gives an overview of how the BlueDot course is structured to empower participants to:
To achieve these goals, courses come in two phases:
The schedule runs from 4 PM UTC Friday to 3 AM Monday. We start with an introductory talk and end the event during the following week with an awards ceremony. Join the public ICal here. You will also find Explorer events, such as collaborative brainstorming and team match-making, before the hackathon begins on Discord and in the calendar.
If you're coming by Taxi or Metro, Enter the Boulevard area of Emirates Towers. You'll find CodersHQ right opposite Creators Hub on the Ground Floor. If you're driving, there is free parking here: https://maps.app.goo.gl/pp1nqVbrN2LiQbk48
If you're coming by Taxi or Metro, Enter the Boulevard area of Emirates Towers. You'll find CodersHQ right opposite Creators Hub on the Ground Floor. If you're driving, there is free parking here: https://maps.app.goo.gl/pp1nqVbrN2LiQbk48
Join us for the Women in AI Safety Research Hackathon in Hereplein 4, 9711GA, Groningen!
Join us for the Women in AI Safety Research Hackathon in Hereplein 4, 9711GA, Groningen!
Join us at the EA Hotel for the Women in AI Safety Research Hackathon. Free accommodation, food and co-working stations provided! We're located at: 36 York Street, Blackpool, FY15AQ. Please register through our Luma event page so we know you're coming!
Join us at the EA Hotel for the Women in AI Safety Research Hackathon. Free accommodation, food and co-working stations provided! We're located at: 36 York Street, Blackpool, FY15AQ. Please register through our Luma event page so we know you're coming!
42AI is a student association dedicated to foster learning and discussion in the field of AI.
42AI is a student association dedicated to foster learning and discussion in the field of AI.
25 Holywell Row, London EC2A 4XE
25 Holywell Row, London EC2A 4XE
Join Warwick AI and Effective Altruism to the joint hackathon Weekend. We will be on the 8th and 9th March in room FAB 2.48 on the main Campus of the Universtiy of Warwick. Anyone can join! Hope to see you there!
Join Warwick AI and Effective Altruism to the joint hackathon Weekend. We will be on the 8th and 9th March in room FAB 2.48 on the main Campus of the Universtiy of Warwick. Anyone can join! Hope to see you there!
Each team should submit a research paper that documents their project and contributions to AI safety. Your submission should demonstrate both technical competence and thoughtful consideration of safety implications within your chosen track.
Your submission package must include the following:
Projects will be evaluated across three main dimensions, with specific considerations for each track:
Q: What is the submission deadline?A: All submissions must be received by March 10, 2025, 3:00 AM UTC.
Q: How many team members are allowed?A: Teams can have 4-5 members. Individual submissions are possible but team collaboration is encouraged.
Q: Can we submit to multiple tracks? A: Teams should focus on one primary track but can incorporate elements from other tracks if relevant.
Q: Do we need to use Goodfire's Ember API?A: Yes, projects in this track should utilize the Ember API for model analysis.
Q: What compute resources are available?A: Each team receives $400 in Lambda Labs credits for compute resources.
Q: Can we build on existing educational platforms?A: Yes, you can integrate with existing platforms while clearly documenting your novel contributions.
Q: How should we measure educational impact?A: Include both quantitative metrics and qualitative assessments of learning outcomes.
Q: What research methodologies are acceptable? A: Both qualitative and quantitative methods are welcome, with clear documentation of methodology.
Q: How should we handle data collection?A: Follow standard social science research ethics and data protection guidelines.
For questions or technical support:
Remember to review all evaluation criteria carefully and ensure your submission addresses the key aspects of your chosen track.