๐ค AI Summary
Balancing adversarial robustness and clean accuracy remains a fundamental challenge in adversarial training. To address this, we propose Memory-based Adversarial Training (MemAT), which introduces a dynamic adversarial example buffer to store and reuse high-value historical adversarial samples across training epochs, coupled with a novel MemLoss objective that jointly optimizes robustness and natural performance. By eliminating redundant adversarial example generation, MemAT significantly improves training efficiency; its buffer scheduling mechanism further enhances sample diversity and effectiveness. On benchmark datasets including CIFAR-10, MemAT achieves state-of-the-art adversarial robustness while boosting clean accuracy by 1.2โ2.8 percentage points on averageโmarking the first demonstration of concurrent gains in both metrics. This work establishes a new paradigm for efficient, high-performance adversarial training.
๐ Abstract
In this paper, we propose a new approach called MemLoss to improve the adversarial training of machine learning models. MemLoss leverages previously generated adversarial examples, referred to as 'Memory Adversarial Examples,' to enhance model robustness and accuracy without compromising performance on clean data. By using these examples across training epochs, MemLoss provides a balanced improvement in both natural accuracy and adversarial robustness. Experimental results on multiple datasets, including CIFAR-10, demonstrate that our method achieves better accuracy compared to existing adversarial training methods while maintaining strong robustness against attacks.