To Mask or to Mirror: Human-AI Alignment in Collective Reasoning

📅 2025-10-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates the alignment between large language models (LLMs) and human social reasoning in collective decision-making, specifically examining whether LLMs reproduce or mitigate human cognitive biases. Method: Leveraging the “Lost at Sea” psychological task, we conduct large-scale human experiments and parallel conditional simulations using Gemini, GPT-4.1, Claude Haiku 3.5, and Gemma 3. A novel group-level human–AI alignment evaluation framework is applied to analyze model reasoning under explicit social cues. Contribution/Results: We identify significant behavioral heterogeneity across models—some replicate human biases, while others actively compensate. Alignment is jointly moderated by situational cues and architectural properties. Critically, this work introduces the first dynamic, collective-scale social alignment assessment paradigm, empirically demonstrating the limitations of static benchmarks. The findings provide both theoretical grounding and empirical evidence for building trustworthy human–AI collaborative decision systems.

Technology Category

Application Category

📝 Abstract
As large language models (LLMs) are increasingly used to model and augment collective decision-making, it is critical to examine their alignment with human social reasoning. We present an empirical framework for assessing collective alignment, in contrast to prior work on the individual level. Using the Lost at Sea social psychology task, we conduct a large-scale online experiment (N=748), randomly assigning groups to leader elections with either visible demographic attributes (e.g. name, gender) or pseudonymous aliases. We then simulate matched LLM groups conditioned on the human data, benchmarking Gemini 2.5, GPT 4.1, Claude Haiku 3.5, and Gemma 3. LLM behaviors diverge: some mirror human biases; others mask these biases and attempt to compensate for them. We empirically demonstrate that human-AI alignment in collective reasoning depends on context, cues, and model-specific inductive biases. Understanding how LLMs align with collective human behavior is critical to advancing socially-aligned AI, and demands dynamic benchmarks that capture the complexities of collective reasoning.
Problem

Research questions and friction points this paper is trying to address.

Examine LLM alignment with human social reasoning
Assess collective alignment beyond individual level analysis
Understand context-dependent biases in AI collective decision-making
Innovation

Methods, ideas, or system contributions that make the work stand out.

Empirical framework assesses collective human-AI alignment
Simulates matched LLM groups using human experimental data
Measures model divergence in mirroring or masking biases
🔎 Similar Papers
No similar papers found.
Crystal Qian
Crystal Qian
Google DeepMind
A
Aaron Parisi
Google DeepMind, San Francisco, USA
C
Clémentine Bouleau
Paris School of Economics, Paris, France
V
Vivian Tsai
Google DeepMind, Mountain View, USA
M
Maël Lebreton
Paris School of Economics, Paris, France
Lucas Dixon
Lucas Dixon
PAIR, Google DeepMind
machine learninginterpretabilitynlpvisualisationgraphical calculi