Apr 19, 2025
Recursive Fitness Alignment Protocol (RFAP)
Andy Williams
Details
Details





The Recursive Fitness Alignment Protocol (RFAP) is a minimal testbed for aligning reasoning systems—AI, human, or institutional—through recursive self-correction. Rather than optimizing for fixed goals or architectures, RFAP begins with the simplest fitness function (a random value between 0 and 1) and introduces a single capability: recursively testing the coherence of reasoning paths. This scaffolds a minimal functional model of intelligence, enabling the identification and correction of misalignment attractors before they become irreversible. RFAP functions as both a theoretical alignment engine and a public experiment designed to stress-test epistemic robustness across diverse reasoning frames.
Cite this work:
@misc {
title={
Recursive Fitness Alignment Protocol (RFAP)
},
author={
Andy Williams
},
date={
4/19/25
},
organization={Apart Research},
note={Research submission to the research sprint hosted by Apart.},
howpublished={https://apartresearch.com}
}
Reviewer's Comments
Reviewer's Comments





No reviews are available yet
Mar 11, 2025
AI Safety Escape Room
The AI Safety Escape Room is an engaging and hands-on AI safety simulation where participants solve real-world AI vulnerabilities through interactive challenges. Instead of learning AI safety through theory, users experience it firsthand – debugging models, detecting adversarial attacks, and refining AI fairness, all within a fun, gamified environment.
Track: Public Education Track
Read More
Mar 10, 2025
Attention Pattern Based Information Flow Visualization Tool
Understanding information flow in transformer-based language models is crucial for mechanistic interpretability. We introduce a visualization tool that extracts and represents attention patterns across model components, revealing how tokens influence each other during processing. Our tool automatically identifies and color-codes functional attention head types based on established taxonomies from recent research on indirect object identification (Wang et al., 2022), factual recall (Chughtai et al., 2024), and factual association retrieval (Geva et al., 2023). This interactive approach enables researchers to trace information propagation through transformer architectures, providing deeper insights into how these models implement reasoning and knowledge retrieval capabilities.
Read More
Mar 10, 2025
LLM Military Decision-Making Under Uncertainty: A Simulation Study
LLMs tested in military decision scenarios typically favor diplomacy over conflict, though uncertainty and chain-of-thought reasoning increase aggressive recommendations. This suggests context-specific limitations for LLM-based military decision support.
Read More
Mar 11, 2025
AI Safety Escape Room
The AI Safety Escape Room is an engaging and hands-on AI safety simulation where participants solve real-world AI vulnerabilities through interactive challenges. Instead of learning AI safety through theory, users experience it firsthand – debugging models, detecting adversarial attacks, and refining AI fairness, all within a fun, gamified environment.
Track: Public Education Track
Read More
Mar 10, 2025
Attention Pattern Based Information Flow Visualization Tool
Understanding information flow in transformer-based language models is crucial for mechanistic interpretability. We introduce a visualization tool that extracts and represents attention patterns across model components, revealing how tokens influence each other during processing. Our tool automatically identifies and color-codes functional attention head types based on established taxonomies from recent research on indirect object identification (Wang et al., 2022), factual recall (Chughtai et al., 2024), and factual association retrieval (Geva et al., 2023). This interactive approach enables researchers to trace information propagation through transformer architectures, providing deeper insights into how these models implement reasoning and knowledge retrieval capabilities.
Read More
Mar 11, 2025
AI Safety Escape Room
The AI Safety Escape Room is an engaging and hands-on AI safety simulation where participants solve real-world AI vulnerabilities through interactive challenges. Instead of learning AI safety through theory, users experience it firsthand – debugging models, detecting adversarial attacks, and refining AI fairness, all within a fun, gamified environment.
Track: Public Education Track
Read More
Mar 10, 2025
Attention Pattern Based Information Flow Visualization Tool
Understanding information flow in transformer-based language models is crucial for mechanistic interpretability. We introduce a visualization tool that extracts and represents attention patterns across model components, revealing how tokens influence each other during processing. Our tool automatically identifies and color-codes functional attention head types based on established taxonomies from recent research on indirect object identification (Wang et al., 2022), factual recall (Chughtai et al., 2024), and factual association retrieval (Geva et al., 2023). This interactive approach enables researchers to trace information propagation through transformer architectures, providing deeper insights into how these models implement reasoning and knowledge retrieval capabilities.
Read More
Mar 11, 2025
AI Safety Escape Room
The AI Safety Escape Room is an engaging and hands-on AI safety simulation where participants solve real-world AI vulnerabilities through interactive challenges. Instead of learning AI safety through theory, users experience it firsthand – debugging models, detecting adversarial attacks, and refining AI fairness, all within a fun, gamified environment.
Track: Public Education Track
Read More
Mar 10, 2025
Attention Pattern Based Information Flow Visualization Tool
Understanding information flow in transformer-based language models is crucial for mechanistic interpretability. We introduce a visualization tool that extracts and represents attention patterns across model components, revealing how tokens influence each other during processing. Our tool automatically identifies and color-codes functional attention head types based on established taxonomies from recent research on indirect object identification (Wang et al., 2022), factual recall (Chughtai et al., 2024), and factual association retrieval (Geva et al., 2023). This interactive approach enables researchers to trace information propagation through transformer architectures, providing deeper insights into how these models implement reasoning and knowledge retrieval capabilities.
Read More

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events

Sign up to stay updated on the
latest news, research, and events