🤖 AI Summary
Stereotypical biases (e.g., gender, race) and structural biases (e.g., lexical overlap, positional preference) in large language models often co-occur due to spurious feature correlations in inputs; existing mitigation methods address them in isolation, risking bias transfer. This paper proposes Causal Contrastive Preference Optimization (CCPO), the first framework unifying both bias types under a shared causal origin—latent-variable-driven spurious correlations. CCPO introduces counterfactual interventions to disentangle bias pathways and designs a fairness-aware, logit-level dynamic attribution and preference update mechanism. Evaluated on nine bias benchmarks—including BBQ and HANS—CCPO significantly reduces bias while preserving model capability, as evidenced by no performance degradation on major competence benchmarks such as MMLU and GSM8K. The approach thus achieves synergistic improvement in both fairness and task performance.
📝 Abstract
Bias in Large Language Models (LLMs) poses significant risks to trustworthiness, manifesting primarily as stereotypical biases (e.g., gender or racial stereotypes) and structural biases (e.g., lexical overlap or position preferences). However, prior paradigms typically address these in isolation, often mitigating one at the expense of exacerbating the other. To address this, we conduct a systematic exploration of these reasoning failures and identify a primary inducement: the latent spurious feature correlations within the input that drive these erroneous reasoning shortcuts. Driven by these findings, we introduce Causal-Contrastive Preference Optimization (C2PO), a unified alignment framework designed to tackle these specific failures by simultaneously discovering and suppressing these correlations directly within the optimization process. Specifically, C2PO leverages causal counterfactual signals to isolate bias-inducing features from valid reasoning paths, and employs a fairness-sensitive preference update mechanism to dynamically evaluate logit-level contributions and suppress shortcut features. Extensive experiments across multiple benchmarks covering stereotypical bias (BBQ, Unqover), structural bias (MNLI, HANS, Chatbot, MT-Bench), out-of-domain fairness (StereoSet, WinoBias), and general utility (MMLU, GSM8K) demonstrate that C2PO effectively mitigates stereotypical and structural biases while preserving robust general reasoning capabilities.