Using Reinforcement Learning to Escape Automatic Filter-based Adversarial Example Defense
Document Type
Article
Publication Date
9-20-2024
Department
Department of Computer Science
Abstract
Deep neural networks can be easily fooled by the adversarial example, which is a specially crafted example with subtle and intentional perturbations. A plethora of papers have proposed to use filters to effectively defend against adversarial example attacks. However, we demonstrate that the automatic filter-based defenses may not be reliable. In this article, we present URL2AED, Using a Reinforcement Learning scheme TO escape the automatic filter-based Adversarial Example Defenses. Specifically, URL2AED uses a specially crafted policy gradient reinforcement learning (RL) algorithm to generate adversarial examples (AEs) that can escape automatic filter-based AE defenses. In particular, we properly design reward functions in policy-gradient RL for targeted attacks and non-targeted attacks, respectively. Furthermore, we customize training algorithms to reduce the possible action space in policy-gradient RL to accelerate URL2AED training while still ensuring that URL2AED generates successful AEs. To demonstrate the performance of the proposed URL2AED, we conduct extensive experiments on three public datasets in terms of different perturbation degrees of parameter, different filter parameters, transferability, and time consumption. The experimental results show that URL2AED achieves high attack success rates for automatic filter-based defenses and good cross-model transferability.
Publication Title
ACM Transactions on Sensor Networks
Recommended Citation
Li, Y.,
Dan, K.,
Lei, X.,
Qin, H.,
Deng, S.,
&
Zhou, G.
(2024).
Using Reinforcement Learning to Escape Automatic Filter-based Adversarial Example Defense.
ACM Transactions on Sensor Networks,
20(5), 1-26.
http://doi.org/10.1145/3688847
Retrieved from: https://digitalcommons.mtu.edu/michigantech-p2/1131