Low-Dimensional Structure in the Space of Language Representations is Reflected in Brain Responses

📅 2021-06-09
🏛️ Neural Information Processing Systems
📈 Citations: 54
Influential: 3
📄 PDF
🤖 AI Summary
This study investigates how representational structures of neural language models, machine translation models, and linguistic annotation tasks relate to each other and to human brain fMRI responses. Method: We systematically analyze representational spaces from 100 models using an integrated framework combining encoder-decoder transfer learning, PCA-based dimensionality reduction, cross-task representation alignment, and neurorepresentational mapping. Contribution/Results: We discover that model representations are embedded in a low-dimensional manifold—introducing the novel concept of “language representation embedding”—and establish the first cross-modal mapping between NLP model hierarchies and cortical language processing hierarchies. The low-dimensional structure significantly predicts mapping fidelity from model representations to fMRI responses; critically, the first principal component robustly recapitulates the cortical hierarchy of language processing (R² = 0.73). This work provides a new paradigm for model interpretability and computational neurolinguistics.
📝 Abstract
How related are the representations learned by neural language models, translation models, and language tagging tasks? We answer this question by adapting an encoder-decoder transfer learning method from computer vision to investigate the structure among 100 different feature spaces extracted from hidden representations of various networks trained on language tasks. This method reveals a low-dimensional structure where language models and translation models smoothly interpolate between word embeddings, syntactic and semantic tasks, and future word embeddings. We call this low-dimensional structure a language representation embedding because it encodes the relationships between representations needed to process language for a variety of NLP tasks. We find that this representation embedding can predict how well each individual feature space maps to human brain responses to natural language stimuli recorded using fMRI. Additionally, we find that the principal dimension of this structure can be used to create a metric which highlights the brain's natural language processing hierarchy. This suggests that the embedding captures some part of the brain's natural language representation structure.
Problem

Research questions and friction points this paper is trying to address.

Analyzes relationships among neural language models and translation models
Investigates low-dimensional structure in language representation spaces
Predicts brain response mapping using representation embeddings
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adapted encoder-decoder transfer learning from vision
Revealed low-dimensional structure among language representations
Embedding predicts brain response mapping and hierarchy
🔎 Similar Papers
No similar papers found.
R
Richard J. Antonello
UT Austin
Javier Turek
Javier Turek
Principal Research Scientist in Machine Learning
Machine LearningNeural NetworksTransformersUnsupervised LearningLLMs
V
Vy A. Vo
Intel Labs
A
Alexander G. Huth
UT Austin