Paying Alignment Tax with Contrastive Learning

📅 2025-05-25
📈 Citations: 0
✨ Influential: 0
📄 PDF
🤖 AI Summary
Existing debiasing methods often induce model capability degradation—manifesting as reduced factual accuracy, knowledge loss, or diminished output readability—posing a fundamental trade-off between capability and fairness, especially in small- and medium-scale models. This paper proposes a contrastive learning–based alignment framework for large language models, the first to simultaneously suppress toxicity and enhance faithfulness. Methodologically, we introduce a dynamic loss scaling mechanism and explicit positive–negative sample contrast to construct controllable, interpretable alignment objectives. Comprehensive evaluation across multiple model scales and diverse benchmarks demonstrates an average 38% reduction in toxicity and a 22% improvement in faithfulness. Critically, both core metrics exhibit consistent improvement across all benchmarks—a first-of-its-kind result—effectively eliminating the “alignment tax.”

Technology Category

Application Category

📝 Abstract
Current debiasing approaches often result a degradation in model capabilities such as factual accuracy and knowledge retention. Through systematic evaluation across multiple benchmarks, we demonstrate that existing debiasing methods face fundamental trade-offs, particularly in smaller models, leading to reduced truthfulness, knowledge loss, or unintelligible outputs. To address these limitations, we propose a contrastive learning framework that learns through carefully constructed positive and negative examples. Our approach introduces contrast computation and dynamic loss scaling to balance bias mitigation with faithfulness preservation. Experimental results across multiple model scales demonstrate that our method achieves substantial improvements in both toxicity reduction and faithfulness preservation. Most importantly, we show that our framework is the first to consistently improve both metrics simultaneously, avoiding the capability degradation characteristic of existing approaches. These results suggest that explicit modeling of both positive and negative examples through contrastive learning could be a promising direction for reducing the alignment tax in language model debiasing.
Problem

Research questions and friction points this paper is trying to address.

Debiasing methods degrade model capabilities like factual accuracy
Existing methods face trade-offs in truthfulness and knowledge retention
Proposing contrastive learning to balance bias mitigation and faithfulness
Innovation

Methods, ideas, or system contributions that make the work stand out.

Contrastive learning with positive and negative examples
Dynamic loss scaling for balance
Simultaneous toxicity reduction and faithfulness preservation
Buse Sibel Korkmaz
Buse Sibel Korkmaz
Imperial College London
R
Rahul Nair
IBM Research Europe, Dublin, Ireland
Elizabeth M. Daly
Elizabeth M. Daly
IBM Research
Interactive AIRecommender SystemsSocial Network Analysis
A
Antonio E. del-Rio Chanona
Imperial College London, London, UK