Loading [MathJax]/extensions/MathMenu.js
Adversarial Example Soups: Improving Transferability and Stealthiness for Free | IEEE Journals & Magazine | IEEE Xplore

Adversarial Example Soups: Improving Transferability and Stealthiness for Free


Abstract:

Transferable adversarial examples cause practical security risks since they can mislead a target model without knowing its internal knowledge. A conventional recipe for m...Show More

Abstract:

Transferable adversarial examples cause practical security risks since they can mislead a target model without knowing its internal knowledge. A conventional recipe for maximizing transferability is to keep only the optimal adversarial example from all those obtained in the optimization pipeline. In this paper, for the first time, we revisit this convention and demonstrate that those discarded, sub-optimal adversarial examples can be reused to boost transferability. Specifically, we propose “Adversarial Example Soups” (AES), with AES-tune for averaging discarded adversarial examples in hyperparameter tuning and AES-rand for stability testing. In addition, our AES is inspired by “model soups”, which averages weights of multiple fine-tuned models for improved accuracy without increasing inference time. Extensive experiments validate the global effectiveness of our AES, boosting 10 state-of-the-art transfer attacks and their combinations by up to 13% against 10 diverse (defensive) target models. We also show the possibility of generalizing AES to other types, e.g., directly averaging multiple in-the-wild adversarial examples that yield comparable success. A promising byproduct of AES is the improved stealthiness of adversarial examples since the perturbation variances are naturally reduced.
Page(s): 1882 - 1894
Date of Publication: 30 January 2025

ISSN Information:

Funding Agency:


I. Introduction

In recent years, Deep Neural Networks (DNNs) have achieved great success in various domains, such as image classification [1], [2], face recognition [3], [4], object detection [5], [6], [7], and autonomous driving [8], [9]. However, DNNs are known to be vulnerable to adversarial examples [10], [11], [12], [13], [14], which are crafted by adding imperceptible perturbations into clean images. Adversarial examples can cause severe threats in black-box security-sensitive applications, such as face recognition systems [15] and autonomous driving cars [16], due to their transferability, i.e., the adversarial examples generated on the surrogate model can be directly used to mislead unknown target models [17], [18], [19], [20], [21].

Contact IEEE to Subscribe

References

References is not available for this document.