BibTeX Entry


@techreport{FrancoEtAl:arxiv26b,
  author	= {Franco, Gabriel and Tassis, Lucas M. and Rohr, Azalea and Crovella, Mark},
  title		= {Finding Highly Interpretable Prompt-Specific Circuits in Language Models},
  year		= {2026},
  URL		= {https://arxiv.org/abs/2602.13483},
  doi		= {10.48550/arXiv.2602.13483},
  abstract	= {Understanding the internal circuits that language models use to solve tasks remains a central challenge in mechanistic interpretability. Most prior work identifies circuits at the \emph{task level} by averaging across many prompts, implicitly assuming a single stable mechanism per task. We show that this assumption can obscure a crucial source of structure: \emph{circuits are prompt-specific}, even within a fixed task. Building on attention causal communication (ACC) \c{i}te{franco2025pinpointing}, we introduce ACC++, refinements that extract cleaner, lower-dimensional \emph{causal} signals inside attention heads from a \emph{single forward pass}. Like ACC, our approach does not require replacement models (e.g., SAEs) or activation patching; ACC++ further improves circuit precision by reducing attribution noise. Applying ACC++ to indirect object identification (IOI) in GPT-2, Pythia, and Gemma 2, we find there is no single circuit for IOI in any model: different prompt templates induce systematically different mechanisms. Despite this variation, prompts cluster into \emph{prompt families} with similar circuits, and we propose a \emph{representative circuit} for each family as a practical unit of analysis. Finally, we develop an automated interpretability pipeline that uses ACC++ signals to surface human-interpretable features and assemble mechanistic explanations for prompt families behavior. Together, our results recast circuits as a meaningful object of study by shifting the unit of analysis from tasks to prompts, enabling scalable circuit descriptions in the presence of prompt-specific mechanisms.}
}