Dynamic Routing and Knowledge Re-Learning for Data-Free Black-Box Attack | IEEE Journals & Magazine | IEEE Xplore

Dynamic Routing and Knowledge Re-Learning for Data-Free Black-Box Attack


Abstract:

Deep learning models have emerged as strong and efficient tools that can be applied to a broad spectrum of complex learning problems and many real-world applications. How...Show More

Abstract:

Deep learning models have emerged as strong and efficient tools that can be applied to a broad spectrum of complex learning problems and many real-world applications. However, more and more works show that deep models are vulnerable to adversarial examples. Compared to vanilla attack settings, this paper advocates a more practical setting of data-free black-box attack, for which the attackers can completely not access the structures and parameters of the target model, as well as the intermediate features and any training data associated with the model. To tackle this task, previous methods generate transferable adversarial examples from a transparent substitute model to the target model. However, we found that these works have the limitations of taking static substitute model structure for different targets, only using hard synthesized examples once, and still relying on data statistics of the target model. This may potentially harm the performance of attacking the target model. To this end, we propose a novel Dynamic Routing and Knowledge Re-Learning framework (DraKe) to effectively learn a dynamic substitute model from the target model. Specifically, given synthesized training samples, a dynamic substitute structure learning strategy is proposed to adaptively generate optimal substitute model structure via a policy network according to different target models and tasks. To facilitate the substitute training, we present a graph-based structure information learning to capture the structural knowledge learned from the target model. For the inherent limitation that online data generation can only be learned once, a dynamic knowledge re-learning strategy is proposed to adjust the weights of optimization objectives and re-learn hard samples. Extensive experiments on four public image classification datasets and one face recognition benchmark are conducted to evaluate the efficacy of our Drake. We can obtain significant improvement compared with state-of-the-art competit...
Page(s): 486 - 501
Date of Publication: 27 September 2024

ISSN Information:

PubMed ID: 39331554

Funding Agency:


I. Introduction

Deep learning has achieved remarkable performance in a wide spectrum of challenging vision applications, such as autonomous driving [1], [2], face recognition [3], [4], person re-identification [5], [6], and computer-aided diagnosis [7], [8]. However, recently some works [9], [10] have shown that the deep models lack robustness and are highly vulnerable to the input adversarial examples. For example, given an input image, an adversarial attack of a target model is formulated as crafting the small perturbations on this image to fool the target model. And this adversarial example will be misclassified with very high confidence. Thus, the adversarial examples reveal important risks in deploying the deep learning models to many real-world applications. Recently, many extraordinary efforts have been made to study the tasks of adversarial attacks and defenses for better assessing and improving the robustness of deep models.

Contact IEEE to Subscribe

References

References is not available for this document.