Nrat: towards adversarial training with inherent label noise

Chen, Zhen and Wang, Fu and Mu, Ronghui and Xu, Peipei and Huang, Xiaowei and Ruan, Wenjie (2024) Nrat: towards adversarial training with inherent label noise. Machine Learning, 113 (6): 6. pp. 3589-3610. ISSN 0885-6125

Full text not available from this repository.

Abstract

Adversarial training (AT) has been widely recognized as the most effective defense approach against adversarial attacks on deep neural networks and it is formulated as a min-max optimization. Most AT algorithms are geared towards research-oriented datasets such as MNIST, CIFAR10, etc., where the labels are generally correct. However, noisy labels, e.g., mislabelling, are inevitable in real-world datasets. In this paper, we investigate AT with inherent label noise, where the training dataset itself contains mislabeled samples. We first empirically show that the performance of AT typically degrades as the label noise rate increases. Then, we propose a Noisy-Robust Adversarial Training (NRAT) algorithm, which leverages the recent advancements in learning with noisy labels to enhance the performance of AT in the presence of label noise. For experimental comparison, we consider two essential metrics in AT: (i) trade-off between natural and robust accuracy; (ii) robust overfitting. Our experiments show that NRAT’s performance is on par with, or better than, the state-of-the-art AT methods on both evaluation metrics. Our code is publicly available at: https://github.com/TrustAI/NRAT.

Item Type:
Journal Article
Journal or Publication Title:
Machine Learning
Uncontrolled Keywords:
/dk/atira/pure/subjectarea/asjc/1700/1702
Subjects:
?? artificial intelligencesoftwareartificial intelligencesoftware ??
ID Code:
235408
Deposited By:
Deposited On:
11 Feb 2026 13:15
Refereed?:
Yes
Published?:
Published
Last Modified:
11 Feb 2026 22:55