Invariance Pair-Guided Learning: Enhancing Robustness in Neural Networks

📅 2025-02-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing out-of-distribution generalization methods suffer from susceptibility to spurious correlations and typically rely on multi-domain labels or hand-crafted group annotations and data augmentation—constraints that limit practical applicability. Method: This paper proposes a robust training framework that requires neither domain/group labels nor specialized data augmentation. Its core innovation is the construction of semantic invariance pairs, from which adaptive corrective gradients are derived and unsupervisedly injected into standard backpropagation to enforce invariance constraints. Contribution/Results: By eliminating dependence on multi-source domain data and explicit grouping supervision, the method enables end-to-end robust learning. Evaluated on ColoredMNIST, Waterbird-100, and CelebA under group shift settings, it achieves an average accuracy improvement of 7.2%, demonstrating significantly enhanced generalization under distributional shifts.

Technology Category

Application Category

📝 Abstract
Out-of-distribution generalization of machine learning models remains challenging since the models are inherently bound to the training data distribution. This especially manifests, when the learned models rely on spurious correlations. Most of the existing approaches apply data manipulation, representation learning, or learning strategies to achieve generalizable models. Unfortunately, these approaches usually require multiple training domains, group labels, specialized augmentation, or pre-processing to reach generalizable models. We propose a novel approach that addresses these limitations by providing a technique to guide the neural network through the training phase. We first establish input pairs, representing the spurious attribute and describing the invariance, a characteristic that should not affect the outcome of the model. Based on these pairs, we form a corrective gradient complementing the traditional gradient descent approach. We further make this correction mechanism adaptive based on a predefined invariance condition. Experiments on ColoredMNIST, Waterbird-100, and CelebA datasets demonstrate the effectiveness of our approach and the robustness to group shifts.
Problem

Research questions and friction points this paper is trying to address.

Enhances out-of-distribution generalization in neural networks.
Addresses reliance on spurious correlations in model training.
Proposes adaptive corrective gradient for robustness to group shifts.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Invariance Pair-Guided Learning
Corrective Gradient Descent
Adaptive Correction Mechanism
🔎 Similar Papers
No similar papers found.
M
Martin Surner
Institute for Data and Process Science, Landshut University of Applied Sciences, Am Lurzenhof 1, Landshut, 84036, Bavaria, Germany; Computer Science Department, Landshut University of Applied Sciences, Am Lurzenhof 1, Landshut, 84036, Bavaria, Germany
A
Abdelmajid Khelil
Institute for Data and Process Science, Landshut University of Applied Sciences, Am Lurzenhof 1, Landshut, 84036, Bavaria, Germany; Computer Science Department, Landshut University of Applied Sciences, Am Lurzenhof 1, Landshut, 84036, Bavaria, Germany
Ludwig Bothmann
Ludwig Bothmann
Postdoctoral Researcher, LMU Munich
StatisticsFairness-aware MLInterpretable MLCausal InferenceComputer Vision