I. Introduction
Deep neural networks (DNNs) have widely deployed in various of daily tasks, e.g., image classification [1], [2], object detection [3], [4], autonomous driving [5], etc. However, DNNs are known to be vulnerable to adversarial examples generated by overlaying carefully designed perturbation into original/natural examples [6], [7], [8]. Against those adversarial examples, the adversarial training is explored to improve the robustness of DNNs, typically by feeding adversarial examples to the model during the training stage [9], [10], [11]. Adversarial training can be formulated as a min-max optimization problem, where perturbation is generated to maximize the original loss, and then the model is optimized against the perturbation/attacks by minimizing the loss [12]. Its goal is to make the model robust and prompt the model to correctly classify the input samples, even with the adversarial perturbation.