FAD: Fine-Grained Adversarial Detection by Perturbation Intensity Classification.
Entropy (Basel)
; 25(2)2023 Feb 11.
Article
in En
| MEDLINE
| ID: mdl-36832701
Adversarial examples present a severe threat to deep neural networks' application in safetycritical domains such as autonomous driving. Although there are numerous defensive solutions, they all have some flaws, such as the fact that they can only defend against adversarial attacks with a limited range of adversarial intensities. Therefore, there is a need for a detection method that can distinguish the adversarial intensity in a fine-grained manner so that subsequent tasks can perform different defense processing against perturbations of various intensities. Based on thefact that adversarial attack samples of different intensities are significantly different in the highfrequency region, this paper proposes a method to amplify the high-frequency component of the image and input it into the deep neural network based on the residual block structure. To our best knowledge, the proposed method is the first to classify adversarial intensities at a fine-grained level, thus providing an attack detection component for a general AI firewall. Experimental results show that our proposed method not only has advanced performance in AutoAttack detection by perturbation intensity classification, but also can effectively apply to detect examples of unseen adversarial attack methods.
Full text:
1
Collection:
01-internacional
Database:
MEDLINE
Type of study:
Diagnostic_studies
Language:
En
Journal:
Entropy (Basel)
Year:
2023
Document type:
Article
Affiliation country:
China
Country of publication:
Switzerland