对抗制
稳健性(进化)
计算机科学
蒸馏
人工智能
机器学习
深层神经网络
人工神经网络
生物化学
基因
有机化学
化学
作者
Shiji Zhao,Jie Yu,Zhenlong Sun,Bo Zhang,Xingxing Wei
标识
DOI:10.1007/978-3-031-19772-7_34
摘要
Adversarial training is an effective approach for improving the robustness of deep neural networks against adversarial attacks. Although bringing reliable robustness, adversarial training (AT) will reduce the performance of identifying clean examples. Meanwhile, Adversarial training can bring more robustness for large models than small models. To improve the robust and clean accuracy of small models, we introduce the Multi-Teacher Adversarial Robustness Distillation (MTARD) to guide the adversarial training process of small models. Specifically, MTARD uses multiple large teacher models, including an adversarial teacher and a clean teacher to guide a small student model in the adversarial training by knowledge distillation. In addition, we design a dynamic training algorithm to balance the influence between the adversarial teacher and clean teacher models. A series of experiments demonstrate that our MTARD can outperform the state-of-the-art adversarial training and distillation methods against various adversarial attacks. Our code is available at https://github.com/zhaoshiji123/MTARD .
科研通智能强力驱动
Strongly Powered by AbleSci AI