SUBSTITUTE MODEL GENERATION FOR BLACK-BOX ADVERSARIAL ATTACK BASED ON KNOWLEDGE DISTILLATION
ï¼·eiyu Cui, Xiaorui Li, Jiawei Huang, Wenyi Wang, Shuai Wang, Jianwen Chen
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 10:38
Although deep convolutional neural network (CNN) performs well in many computer vision tasks, its classification mechanism is very vulnerable when it is exposed to the perturbation of adversarial attacks. In this paper, we proposed a new algorithm to generate the substitute model of black-box CNN models by using knowledge distillation. The proposed algorithm distills multiple CNN teacher models to a compact student model as the substitution of other black-box CNN models to be attacked. The black-box adversarial samples can be consequently generated on this substitute model by using various white-box attacking methods. According to our experiments on ResNet18 and DenseNet121, our algorithm boosts the attacking success rate (ASR) by 20% by training the substitute model based on knowledge distillation.