Adversarial Instance Generation and Robust Training for Neural Combinatorial Optimization with Multiple Objectives

📅 2026-01-04
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing neural solvers for multi-objective combinatorial optimization problems often lack robustness on out-of-distribution and complex instances. This work proposes the first preference-guided adversarial attack and defense framework tailored to such problems. The approach generates challenging samples through preference-conditioned adversarial attacks and enhances model generalization via adversarial training integrated with a hardness-aware preference selection mechanism. Evaluated on standard benchmarks—including the Multi-Objective Traveling Salesman Problem (MOTSP), Multi-Objective Capacitated Vehicle Routing Problem (MOCVRP), and Multi-Objective Knapsack Problem (MOKP)—the method demonstrates significant improvements in both performance and stability under difficult and out-of-distribution scenarios.

Technology Category

Application Category

📝 Abstract
Deep reinforcement learning (DRL) has shown great promise in addressing multi-objective combinatorial optimization problems (MOCOPs). Nevertheless, the robustness of these learning-based solvers has remained insufficiently explored, especially across diverse and complex problem distributions. In this paper, we propose a unified robustness-oriented framework for preference-conditioned DRL solvers for MOCOPs. Within this framework, we develop a preference-based adversarial attack to generate hard instances that expose solver weaknesses, and quantify the attack impact by the resulting degradation on Pareto-front quality. We further introduce a defense strategy that integrates hardness-aware preference selection into adversarial training to reduce overfitting to restricted preference regions and improve out-of-distribution performance. The experimental results on multi-objective traveling salesman problem (MOTSP), multi-objective capacitated vehicle routing problem (MOCVRP), and multi-objective knapsack problem (MOKP) verify that our attack method successfully learns hard instances for different solvers. Furthermore, our defense method significantly strengthens the robustness and generalizability of neural solvers, delivering superior performance on hard or out-of-distribution instances.
Problem

Research questions and friction points this paper is trying to address.

multi-objective combinatorial optimization
robustness
deep reinforcement learning
adversarial instances
out-of-distribution generalization
Innovation

Methods, ideas, or system contributions that make the work stand out.

adversarial instance generation
robust training
multi-objective combinatorial optimization
preference-conditioned DRL
hardness-aware preference selection
🔎 Similar Papers
No similar papers found.
W
Wei Liu
Leiden Institute of Advanced Computer Science, Leiden University, Leiden, The Netherlands
Yaoxin Wu
Yaoxin Wu
Eindhoven University of Technology
Deep learningCombinatorial optimizationInteger programmingMulti-objective optimization
Yingqian Zhang
Yingqian Zhang
Associate Professor of AI for Decision-Making, Eindhoven University of Technology
Artificial IntelligenceData-Driven OptimizationDeep RLSocial-aware Algorithms
T
Thomas Back
Leiden Institute of Advanced Computer Science, Leiden University, Leiden, The Netherlands
Y
Yingjie Fan
Leiden Institute of Advanced Computer Science, Leiden University, Leiden, The Netherlands