Seeking Consistent Flat Minima for Better Domain Generalization via Refining Loss Landscapes

πŸ“… 2024-12-18
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing domain generalization methods overlook cross-domain consistency of loss landscapes when optimizing flat minima, hindering simultaneous convergence to optimal flat solutions across all source domains and limiting generalization performance. This paper proposes the Self-Feedback Training (SFT) frameworkβ€”an iterative training paradigm that introduces, for the first time, a cross-domain loss landscape consistency constraint. By dynamically detecting and correcting inter-domain landscape inconsistencies, SFT jointly optimizes multi-source domain loss surfaces to achieve cross-domain alignment of flat minima. The method integrates a landscape consistency metric, gradient-driven landscape smoothing regularization, and an extension of Sharpness-Aware Minimization (SAM). Evaluated on five DomainBed benchmarks, SFT with ResNet-50 and ViT-B/16 backbones achieves average improvements of 2.6% and 1.5% over SAM, respectively, demonstrating significantly enhanced out-of-distribution generalization capability.

Technology Category

Application Category

πŸ“ Abstract
Domain generalization aims to learn a model from multiple training domains and generalize it to unseen test domains. Recent theory has shown that seeking the deep models, whose parameters lie in the flat minima of the loss landscape, can significantly reduce the out-of-domain generalization error. However, existing methods often neglect the consistency of loss landscapes in different domains, resulting in models that are not simultaneously in the optimal flat minima in all domains, which limits their generalization ability. To address this issue, this paper proposes an iterative Self-Feedback Training (SFT) framework to seek consistent flat minima that are shared across different domains by progressively refining loss landscapes during training. It alternatively generates a feedback signal by measuring the inconsistency of loss landscapes in different domains and refines these loss landscapes for greater consistency using this feedback signal. Benefiting from the consistency of the flat minima within these refined loss landscapes, our SFT helps achieve better out-of-domain generalization. Extensive experiments on DomainBed demonstrate superior performances of SFT when compared to state-of-the-art sharpness-aware methods and other prevalent DG baselines. On average across five DG benchmarks, SFT surpasses the sharpness-aware minimization by 2.6% with ResNet-50 and 1.5% with ViT-B/16, respectively. The code will be available soon.
Problem

Research questions and friction points this paper is trying to address.

Seeking consistent flat minima across domains
Refining loss landscapes for better generalization
Reducing inconsistency in multi-domain loss landscapes
Innovation

Methods, ideas, or system contributions that make the work stand out.

Self-Feedback Training for consistent flat minima
Refining loss landscapes across domains iteratively
Measuring and reducing loss landscape inconsistency
πŸ”Ž Similar Papers
No similar papers found.
A
Aodi Li
University of Science and Technology of China, Hefei 230026, China
Liansheng Zhuang
Liansheng Zhuang
Univerisity of Science and Technology of China
Computer VisionKnowledge GraphComputer Games
Xiao Long
Xiao Long
University of Science and Technology of China | Alibaba Group
Knowledge GraphLarge Language ModelReasoning
M
Minghong Yao
University of Science and Technology of China, Hefei 230026, China
S
Shafei Wang
Peng Cheng Laboratory, Shenzhen 518000, China