Geometry of Decision Making in Language Models

📅 2025-11-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) exhibit opaque decision-making processes, hindering interpretability and mechanistic understanding. Method: We systematically analyze the geometric evolution of hidden representations across layers of 28 open-source Transformer models on multiple-choice question answering (MCQA), quantifying intrinsic dimensionality (ID) layerwise using multiple ID estimators. Results: We identify a consistent three-phase ID dynamic—“low-dimensional → expansion → compression”—across models and datasets: early layers rapidly project inputs onto task-relevant low-dimensional manifolds; middle layers expand representations to support reasoning; late layers compress them into discriminative subspaces. This universal geometric pattern provides the first empirical evidence of a shared structural basis underlying LLM generalization and emergent reasoning capabilities. It establishes a verifiable, structured interpretability framework for probing implicit decision mechanisms in LLMs.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) show strong generalization across diverse tasks, yet the internal decision-making processes behind their predictions remain opaque. In this work, we study the geometry of hidden representations in LLMs through the lens of extit{intrinsic dimension} (ID), focusing specifically on decision-making dynamics in a multiple-choice question answering (MCQA) setting. We perform a large-scale study, with 28 open-weight transformer models and estimate ID across layers using multiple estimators, while also quantifying per-layer performance on MCQA tasks. Our findings reveal a consistent ID pattern across models: early layers operate on low-dimensional manifolds, middle layers expand this space, and later layers compress it again, converging to decision-relevant representations. Together, these results suggest LLMs implicitly learn to project linguistic inputs onto structured, low-dimensional manifolds aligned with task-specific decisions, providing new geometric insights into how generalization and reasoning emerge in language models.
Problem

Research questions and friction points this paper is trying to address.

Analyzing geometric patterns in LLM representations during decision-making processes
Investigating intrinsic dimension evolution across transformer layers in MCQA tasks
Understanding how linguistic inputs project onto task-aligned low-dimensional manifolds
Innovation

Methods, ideas, or system contributions that make the work stand out.

Intrinsic dimension analysis reveals hidden representation geometry
Early layers use low-dimensional manifolds, middle layers expand space
Later layers compress representations to align with decisions
🔎 Similar Papers
A
Abhinav Joshi
Indian Institute of Technology Kanpur (IIT Kanpur)
D
Divyanshu Bhatt
Samsung Research & Development Institute, Bangalore
Ashutosh Modi
Ashutosh Modi
Indian Institute of Technology Kanpur
Natural Language ProcessingMachine and Deep LearningArtificial IntelligenceAffective ComputingLegal AI