A Second-Order Perspective on Model Compositionality and Incremental Learning

📅 2024-05-25
📈 Citations: 4
Influential: 0
📄 PDF
🤖 AI Summary
This work investigates the intrinsic conditions under which deep pretrained models can achieve modular compositionality in nonlinear networks, enabling multi-task modeling, task specialization, and selective forgetting. Method: We propose a theoretical framework based on the second-order Taylor expansion of the loss function—introducing, for the first time, a second-order optimization perspective into compositional analysis—and identify “preserving the pretrained attractor basin” as the key mechanism ensuring modular composability. Building on this insight, we design a dual-path incremental training algorithm that supports module reuse, dynamic task integration, and controllable knowledge卸载 (unloading). Results: Empirical evaluation on incremental classification tasks demonstrates that the constructed module pool enables efficient synthesis of high-performance multi-task models, improves single-task accuracy, and facilitates targeted knowledge erasure—thereby significantly enhancing model editability and adaptability.

Technology Category

Application Category

📝 Abstract
The fine-tuning of deep pre-trained models has revealed compositional properties, with multiple specialized modules that can be arbitrarily composed into a single, multi-task model. However, identifying the conditions that promote compositionality remains an open issue, with recent efforts concentrating mainly on linearized networks. We conduct a theoretical study that attempts to demystify compositionality in standard non-linear networks through the second-order Taylor approximation of the loss function. The proposed formulation highlights the importance of staying within the pre-training basin to achieve composable modules. Moreover, it provides the basis for two dual incremental training algorithms: the one from the perspective of multiple models trained individually, while the other aims to optimize the composed model as a whole. We probe their application in incremental classification tasks and highlight some valuable skills. In fact, the pool of incrementally learned modules not only supports the creation of an effective multi-task model but also enables unlearning and specialization in certain tasks. Code available at https://github.com/aimagelab/mammoth.
Problem

Research questions and friction points this paper is trying to address.

Understanding conditions promoting model compositionality in non-linear networks.
Developing incremental training algorithms for multi-task model optimization.
Exploring unlearning and specialization in incrementally learned modules.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Second-order Taylor approximation for compositionality analysis
Dual incremental training algorithms for model optimization
Incremental learning supports multi-task model creation
🔎 Similar Papers
No similar papers found.
Angelo Porrello
Angelo Porrello
Università di Modena e Reggio Emilia
Artificial IntelligenceMachine LearningDeep LearningContinual LearningModular Deep Learning
Lorenzo Bonicelli
Lorenzo Bonicelli
Symboolic S.r.l.
Machine LearningDeep Learning
P
Pietro Buzzega
AImageLab - University of Modena and Reggio Emilia, Modena, Italy
M
Monica Millunzi
AImageLab - University of Modena and Reggio Emilia, Modena, Italy
Simone Calderara
Simone Calderara
University of Modena and Reggio Emilia
Machine learningcontinual learningtrackingpattern recognition
R
R. Cucchiara
AImageLab - University of Modena and Reggio Emilia, Modena, Italy