🤖 AI Summary
Large language models (LLMs) often perpetuate and amplify harmful societal biases—such as gender and racial stereotypes—posing critical safety risks in real-world applications. To address the limitations of existing approaches, which rely on opaque data filtering or post-hoc debiasing, this paper introduces the first end-to-end interpretable bias intervention framework. It employs linear probing to precisely localize bias representations within the model’s internal activations and dynamically suppresses bias-correlated neural activity during inference via activation steering vectors. Evaluated on GPT2-large, the method achieves near-perfect bias detection accuracy (~100%) and significantly reduces stereotypical content in generated text while improving output neutrality. This work pioneers the tight integration of interpretable probing with real-time activation-level intervention, jointly ensuring transparency in bias identification and controllability in mitigation—establishing a novel paradigm for safe and aligned LLM deployment.
📝 Abstract
As large language models (LLMs) become more integrated into societal systems, the risk of them perpetuating and amplifying harmful biases becomes a critical safety concern. Traditional methods for mitigating bias often rely on data filtering or post-hoc output moderation, which treat the model as an opaque black box. In this work, we introduce a complete, end-to-end system that uses techniques from mechanistic interpretability to both identify and actively mitigate bias directly within a model's internal workings. Our method involves two primary stages. First, we train linear "probes" on the internal activations of a model to detect the latent representations of various biases (e.g., gender, race, age). Our experiments on exttt{gpt2-large} demonstrate that these probes can identify biased content with near-perfect accuracy, revealing that bias representations become most salient in the model's later layers. Second, we leverage these findings to compute "steering vectors" by contrasting the model's activation patterns for biased and neutral statements. By adding these vectors during inference, we can actively steer the model's generative process away from producing harmful, stereotypical, or biased content in real-time. We demonstrate the efficacy of this activation steering technique, showing that it successfully alters biased completions toward more neutral alternatives. We present our work as a robust and reproducible system that offers a more direct and interpretable approach to building safer and more accountable LLMs.