Efficient Semi-Supervised Adversarial Training via Latent Clustering-Based Data Reduction

Abstract

Learning robust models under adversarial settings is widely recognized as requiring a considerably large number of training samples. Recent work proposes semi-supervised adversarial training (SSAT), which utilizes external unlabeled or synthetically generated data and is currently the state of the art. However, SSAT requires substantial extra data to attain high robustness, resulting in prolonged training time and increased memory usage. In this paper, we propose data reduction strategies to improve the efficiency of SSAT by optimizing the amount of additional data incorporated. Specifically, we design novel latent clustering-based techniques to select or generate a small, critical subset of data samples near the model’s decision boundary. While focusing on boundary-adjacent points, our methods maintain a balanced ratio between boundary and non-boundary data points, thereby avoiding overfitting. Comprehensive experiments across image benchmarks demonstrate that our methods can effectively reduce SSAT’s data requirements and computational costs while preserving its strong robustness advantages. In particular, our latent-space selection scheme based on k-means clustering and our guided diffusion-based approach with LCG-KM are the most effective, achieving nearly identical robust accuracies with 5x to 10x less unlabeled data. When compared to full SSAT trained to convergence, our methods reduce total runtime by approximately 3x to 4x due to strategic prioritization of unlabeled data.

Publication
The 4th IEEE Conference on Secure and Trustworthy Machine Learning

A preliminary version of this work was presented at NextGenAISafety Workshop at ICML 2024. The workshop paper can be found on Openreview.