Set a Thief to Catch a Thief: Combating Label Noise through Noisy Meta Learning

📅 2025-02-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses learning with noisy labels in the absence of a clean validation set. We propose STCT, a meta-learning framework that requires no additional clean data. Its core innovation is the “noise-cleans-noise” paradigm: it treats noise-distributed training data itself as a meta-validation set, enabling decoupled optimization of noise-aware meta-correction and semi-supervised representation learning via alternating updates. STCT integrates noisy meta-learning, dynamic label refinement, consistency regularization, and semi-supervised learning. On CIFAR-10 under 80% symmetric label noise, it achieves 96.9% label correction accuracy and 95.2% classification accuracy—surpassing state-of-the-art methods. To our knowledge, STCT is the first approach to achieve highly robust noise label correction and model training without relying on any clean validation data.

Technology Category

Application Category

📝 Abstract
Learning from noisy labels (LNL) aims to train high-performance deep models using noisy datasets. Meta learning based label correction methods have demonstrated remarkable performance in LNL by designing various meta label rectification tasks. However, extra clean validation set is a prerequisite for these methods to perform label correction, requiring extra labor and greatly limiting their practicality. To tackle this issue, we propose a novel noisy meta label correction framework STCT, which counterintuitively uses noisy data to correct label noise, borrowing the spirit in the saying ``Set a Thief to Catch a Thief''. The core idea of STCT is to leverage noisy data which is i.i.d. with the training data as a validation set to evaluate model performance and perform label correction in a meta learning framework, eliminating the need for extra clean data. By decoupling the complex bi-level optimization in meta learning into representation learning and label correction, STCT is solved through an alternating training strategy between noisy meta correction and semi-supervised representation learning. Extensive experiments on synthetic and real-world datasets demonstrate the outstanding performance of STCT, particularly in high noise rate scenarios. STCT achieves 96.9% label correction and 95.2% classification performance on CIFAR-10 with 80% symmetric noise, significantly surpassing the current state-of-the-art.
Problem

Research questions and friction points this paper is trying to address.

Eliminates need for clean validation data
Corrects noisy labels using noisy data
Improves performance in high noise scenarios
Innovation

Methods, ideas, or system contributions that make the work stand out.

Noisy data for label correction
Decoupled bi-level optimization strategy
Alternating training for meta correction
🔎 Similar Papers
No similar papers found.
H
Hanxuan Wang
Department of Automation Science and Engineering, Xi’an Jiaotong University, China
Na Lu
Na Lu
Professor of System Engineering Institute, Xi'an Jiaotong University
image analysismachine learningbrain computer interfacerobotics
X
Xueying Zhao
School of Electrical Engineering, Xi’an Jiaotong University, China
Y
Yuxuan Yan
Department of Automation Science and Engineering, Xi’an Jiaotong University, China
K
Kaipeng Ma
Department of Automation Science and Engineering, Xi’an Jiaotong University, China
K
K. C. Keong
College of Computing and Data Science, Nanyang Technological University, Singapore
Gustavo Carneiro
Gustavo Carneiro
Professor of AI and Machine Learning, University of Surrey
Computer VisionMedical Image AnalysisMachine LearningMedical Image Computing