Improving Multi-Label Contrastive Learning by Leveraging Label Distribution

📅 2025-01-31
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenges of complex positive/negative sample selection and unmodeled label importance in multi-label contrastive learning, this paper proposes a simplified sample selection method grounded in label distribution modeling. It is the first to incorporate label distributions into multi-label contrastive learning frameworks, jointly reconstructing soft label distributions via RBF kernel mapping and contrastive loss—thereby explicitly capturing inter-label semantic relationships and enabling importance-aware representation learning. Crucially, the method obviates computationally expensive hard-negative mining and instead adaptively identifies positive/negative sample intersections solely from label co-occurrence structures. Extensive experiments across nine mainstream multi-label datasets—spanning both image and vector modalities—demonstrate consistent superiority over state-of-the-art methods. The approach achieves significant improvements across six core metrics, including mAP and Hamming Loss, validating its effectiveness, robustness, and cross-modal generalizability.

Technology Category

Application Category

📝 Abstract
In multi-label learning, leveraging contrastive learning to learn better representations faces a key challenge: selecting positive and negative samples and effectively utilizing label information. Previous studies selected positive and negative samples based on the overlap between labels and used them for label-wise loss balancing. However, these methods suffer from a complex selection process and fail to account for the varying importance of different labels. To address these problems, we propose a novel method that improves multi-label contrastive learning through label distribution. Specifically, when selecting positive and negative samples, we only need to consider whether there is an intersection between labels. To model the relationships between labels, we introduce two methods to recover label distributions from logical labels, based on Radial Basis Function (RBF) and contrastive loss, respectively. We evaluate our method on nine widely used multi-label datasets, including image and vector datasets. The results demonstrate that our method outperforms state-of-the-art methods in six evaluation metrics.
Problem

Research questions and friction points this paper is trying to address.

Multi-label Learning
Contrastive Learning
Label Importance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Label Distribution
Multi-label Contrastive Learning
Radial Basis Function
🔎 Similar Papers
No similar papers found.
N
Ning Chen
Key Laboratory of Water Big Data Technology of Ministry of Water Resources, College of Computer Science and Software Engineering, Hohai University, Nanjing, China
Shen-Huan Lyu
Shen-Huan Lyu
Hohai University
Artificial IntelligenceMachine LearningData Mining
T
Tian-Shuang Wu
Key Laboratory of Water Big Data Technology of Ministry of Water Resources, College of Computer Science and Software Engineering, Hohai University, Nanjing, China
Y
Yanyan Wang
Key Laboratory of Water Big Data Technology of Ministry of Water Resources, College of Computer Science and Software Engineering, Hohai University, Nanjing, China
B
Bin Tang
Key Laboratory of Water Big Data Technology of Ministry of Water Resources, College of Computer Science and Software Engineering, Hohai University, Nanjing, China