Stream Aligner: Efficient Sentence-Level Alignment via Distribution Induction

📅 2025-01-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) suffer from poor cross-task adaptability, low real-time responsiveness, and heavy reliance on auxiliary large models in value alignment. To address these challenges, we propose the Streaming Distribution Induce Aligner (Stream Aligner), the first framework to introduce *streaming distribution-induced alignment*—a paradigm enabling sentence-level, low-latency, embedded dynamic alignment during LLM autoregressive generation. Stream Aligner employs a lightweight preference model to infer user intent and iteratively refines suffix tokens without invoking external large models. Experiments demonstrate that Stream Aligner-2B improves helpfulness by 76.1% and harmlessness by 36.0% for Llama2-70B-chat; Stream Aligner-8B enhances mathematical reasoning performance by 3.5% for Llama3-70B-Instruct. Our approach significantly advances alignment efficiency and generalization across diverse tasks and model scales.

Technology Category

Application Category

📝 Abstract
The rapid advancement of large language models (LLMs) has led to significant improvements in their capabilities, but also to increased concerns about their alignment with human values and intentions. Current alignment strategies, including adaptive training and inference-time methods, have demonstrated potential in this area. However, these approaches still struggle to balance deployment complexity and capability across various tasks and difficulties. In this work, we introduce the Streaming Distribution Induce Aligner (Stream Aligner), a novel alignment paradigm that combines efficiency with enhanced performance in various tasks throughout the generation process. Stream Aligner achieves dynamic sentence-level correction by using a small model to learn the preferences of the suffix sentence, iteratively correcting the suffix sentence output by the upstream model, and then using the corrected sentence to replace the suffix sentence in subsequent generations. Compared to Aligner, our experiments demonstrate that Stream Aligner reduces reliance on the capabilities of additional models, enhances the reasoning abilities of LLMs, and decreases latency during user interaction. Specifically, Stream Aligner-2B model has achieved an improvement of 76.1% in helpfulness, 36.0% in harmlessness on the tested Llama2-70B-chat model, and Stream Aligner-8B has achieved an improvement of 3.5% on the math ability of the tested Llama3-70B-Instruct model.
Problem

Research questions and friction points this paper is trying to address.

Large Language Models
Human Values Alignment
Task Adaptability and Efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Stream Aligner
Language Model Efficiency
Human Value Alignment
🔎 Similar Papers
No similar papers found.
Hantao Lou
Hantao Lou
Peking University
AI AlignmentAI SafetyInterpretabilityTrustworthy AI
J
Jiaming Ji
Institute for AI, Peking University; State Key Laboratory of General Artificial Intelligence, Institute for AI, Peking University
Kaile Wang
Kaile Wang
Peking University
Y
Yaodong Yang
Institute for AI, Peking University; State Key Laboratory of General Artificial Intelligence, Institute for AI, Peking University