Advancing Supervised Local Learning Beyond Classification with Long-term Feature Bank

📅 2024-06-01
📈 Citations: 3
Influential: 0
📄 PDF
🤖 AI Summary
Existing local learning methods are predominantly confined to image classification, exhibiting limited adaptability to diverse vision tasks such as object detection and super-resolution. Key limitations include insufficient cross-scale feature communication and poor knowledge transfer across tasks. This work pioneers the extension of *supervised* local learning beyond classification to general vision tasks. We propose a Memory-Augmented Auxiliary Network (MAN), integrating a long-term feature bank with a lightweight cross-scale fusion mechanism. MAN enables multi-task compatibility while substantially reducing GPU memory consumption. Evaluated on multiple standard benchmarks—including COCO, PASCAL VOC, and DIV2K—our approach achieves performance on par with end-to-end models, yet with significantly lower memory overhead. These results empirically validate the scalability and effectiveness of the local learning paradigm for broad vision applications.

Technology Category

Application Category

📝 Abstract
Local learning offers an alternative to traditional end-to-end back-propagation in deep neural networks, significantly reducing GPU memory usage. While local learning has shown promise in image classification tasks, its application to other visual tasks remains limited. This limitation arises primarily from two factors: 1) architectures tailored for classification are often not transferable to other tasks, leading to a lack of reusability of task-specific knowledge; 2) the absence of cross-scale feature communication results in degraded performance in tasks such as object detection and super-resolution. To address these challenges, we propose the Memory-augmented Auxiliary Network (MAN), which introduces a simplified design principle and incorporates a feature bank to enhance cross-task adaptability and communication. This work represents the first successful application of local learning methods beyond classification, demonstrating that MAN not only conserves GPU memory but also achieves performance on par with end-to-end approaches across multiple datasets for various visual tasks.
Problem

Research questions and friction points this paper is trying to address.

Extend local learning beyond classification tasks
Overcome limited cross-task knowledge reuse
Address lack of cross-scale feature communication
Innovation

Methods, ideas, or system contributions that make the work stand out.

Feature Bank Augmented auxiliary network design
Enhanced cross-task adaptability and communication
Reduced GPU memory with competitive performance
🔎 Similar Papers
No similar papers found.
F
Feiyu Zhu
University of Shanghai for Science and Technology
Yuming Zhang
Yuming Zhang
University of Kentucky
weldsensormodelingcontrolrobot
Changpeng Cai
Changpeng Cai
Zhipu AI (Z.ai)
Artificial IntelligenceLarge Language ModelMultimodal
G
Guinan Guo
Sun Yat-sen University
Jiao Li
Jiao Li
Columbia University
Applied MathMachine LearningFinanceClimate Change
X
Xiuyuan Guo
University of Southern California
Q
Quanwei Zhang
Zhejiang University
P
Peizhe Wang
Southeast University
C
Chenghao He
East China University of Science and Technology
Junhao Su
Junhao Su
MeiTuan Inc.
Computer Vision