Netinfo Security ›› 2022, Vol. 22 ›› Issue (3): 70-77.doi: 10.3969/j.issn.1671-1122.2022.03.008

Previous Articles     Next Articles

A Defense Method against Adversarial Attacks Based on Neural Architecture Search

ZHENG Yaohao1,2, WANG Liming1(), YANG Jing1   

  1. 1. Institute of Information Engineering, Chinese Academy of Sciences, Beijing 100093, China
    2. School of Cyber Security, University of Chinese Academy of Sciences, Beijing 100049, China
  • Received:2021-09-19 Online:2022-03-10 Published:2022-03-28
  • Contact: WANG Liming E-mail:wangliming@iie.ac.cn

Abstract:

Aiming at the problem that the neural networks are easy to misclassify under the attack of adversarial examples in the task of image classification, which leads to the unreliability of deep learning models, this paper proposed a defense method against adversarial attacks based on neural architecture search. This method used reinforcement learning to model the search of defense network as the behavior of the agent. Through the definition of search space, the design of search strategy, and the evaluation of subnetwork performance, the search network can automatically obtain the best performance network to reconstruct adversarial images and restore them to natural images, achieving the purpose of defense against adversarial attacks. The experimental results show that the method can effectively reconstruct illegal examples, and make them lose aggressiveness, and consequently ensure the classification accuracy of the classifier.

Key words: neural architecture search, image classification, adversarial attack, deep learning

CLC Number: