đ¤ AI Summary
Language models can implicitly encode behavioral traitsâsuch as preferences or biasesâthrough semantically irrelevant data (e.g., numeric sequences, code, reasoning traces), a phenomenon termed âsubconscious learning.â Its universality and cross-architecture transferability remain unclear.
Method: The authors propose a teacherâstudent distillation framework, training student models exclusively on data rigorously stripped of explicit semantic cues. They combine theoretical analysis with MLP-based experiments to isolate the effect of structural correspondence between teacher and student architectures.
Contribution/Results: Even when inputs contain no task-relevant semantics, students faithfully reproduce teachersâ behavioral tendenciesâdemonstrating subconscious learning is intrinsic to neural representations. This effect persists across diverse foundation models but critically depends on low-level architectural alignment between teacher and student. The study provides the first empirical and theoretical evidence that subconscious learning is a pervasive mechanism in neural networks, robust to conventional data sanitization. These findings carry critical implications for AI safety, model distillation, and alignmentâhighlighting inherent limitations of input-level mitigation strategies.
đ Abstract
We study subliminal learning, a surprising phenomenon where language models transmit behavioral traits via semantically unrelated data. In our main experiments, a "teacher" model with some trait T (such as liking owls or being misaligned) generates a dataset consisting solely of number sequences. Remarkably, a "student" model trained on this dataset learns T. This occurs even when the data is filtered to remove references to T. We observe the same effect when training on code or reasoning traces generated by the same teacher model. However, we do not observe the effect when the teacher and student have different base models. To help explain our findings, we prove a theoretical result showing that subliminal learning occurs in all neural networks under certain conditions, and demonstrate subliminal learning in a simple MLP classifier. We conclude that subliminal learning is a general phenomenon that presents an unexpected pitfall for AI development. Distillation could propagate unintended traits, even when developers try to prevent this via data filtering.