A Retention-Centric Framework for Continual Learning with Guaranteed Model Developmental Safety

📅 2024-10-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In cost-sensitive scenarios involving continual model iteration, catastrophic forgetting undermines the reliability of previously acquired capabilities. Method: This paper proposes a novel paradigm—“model development safety”—requiring new models to not only improve performance on target tasks but also strictly preserve critical capabilities of prior models. To this end, we formulate a capability-preserving continual learning framework, introducing for the first time a data-dependent capability constraint mechanism and an efficient constrained optimization algorithm with theoretical guarantees—overcoming the limitations of conventional average-performance trade-offs. Leveraging the CLIP pre-trained model, we adopt task-specific head fine-tuning and jointly optimize the target-task loss and capability preservation constraints. Results: Experiments on autonomous driving and scene recognition benchmarks demonstrate zero degradation in accuracy on key legacy tasks, significant gains on new tasks, and complete prevention of safety-critical capability deterioration.

Technology Category

Application Category

📝 Abstract
In real-world applications, learning-enabled systems often undergo iterative model development to address challenging or emerging tasks, which involve collecting new data, training a new model and validating the model. This continual model development process raises a significant issue that acquiring new or improving existing capabilities may inadvertently lose good capabilities of the old model, also known as catastrophic forgetting. While existing continual learning aims to mitigate catastrophic forgetting by trading off performance on previous tasks and new tasks to ensure good average performance, it often falls short in cost-sensitive applications, where failing to preserve essential established capabilities introduces unforeseen costs and risks and substantial expenses for re-improving these capabilities. To address this issue, we impose a requirement on learning systems to ensure that a new model strictly retains important capabilities of the old model while improving target-task performance, which we term model developmental safety. To ensure model developmental safety, we propose a retention-centric framework with data-dependent constraints, and study how to continually develop a pretrained CLIP model for acquiring new or improving existing capabilities of image classification. We propose an efficient constrained optimization algorithm with theoretical guarantees and use its insights to finetune the CLIP model with task-dependent heads for promoting the model developmental safety. Experiments on autonomous driving and scene recognition datasets validate the efficacy of our method.
Problem

Research questions and friction points this paper is trying to address.

Ensures new models retain old model's essential capabilities
Addresses catastrophic forgetting in continual learning systems
Improves target-task performance with developmental safety
Innovation

Methods, ideas, or system contributions that make the work stand out.

Retention-centric framework for continual learning
Data-dependent constraints ensure model safety
Efficient constrained optimization algorithm
🔎 Similar Papers
No similar papers found.
G
Gang Li
Department of Computer Science & Engineering, Texas A&M University, College Station, USA.
Wendi Yu
Wendi Yu
Texas A&M University
Machine Learning
Y
Yao Yao
Department of Mathematics, The University of Iowa, Iowa City, USA.
Wei Tong
Wei Tong
Global Research and Development, General Motors, Warren, USA.
Y
Yingbin Liang
Department of Electrical and Computer Engineering, The Ohio State University, Columbus, UAS.
Qihang Lin
Qihang Lin
The University of Iowa
Continuous optimizationStochastic OptimizationMachine LearningMarkov Decision Process
Tianbao Yang
Tianbao Yang
Texas A&M University
machine learningstochastic optimization