稳健性(进化)
对抗制
计算机科学
人工智能
机器学习
深层神经网络
数据挖掘
深度学习
生物化学
化学
基因
作者
Hamid Eghbal-zadeh,Werner Zellinger,Maura Pintor,Kathrin Grosse,Khaled Koutini,Bernhard Moser,Battista Biggio,Gerhard Widmer
标识
DOI:10.1016/j.ins.2023.119838
摘要
Recent work has proposed novel data augmentation methods to improve the adversarial robustness of deep neural networks. In this paper, we re-evaluate such methods through the lens of different metrics that characterize the augmented manifold, finding contradictory evidence. Our extensive empirical analysis involving 5 data augmentation methods, all tested with an increasing probability of augmentation, shows that: (i) novel data augmentation methods proposed to improve adversarial robustness only improve it when combined with classical augmentations (like image flipping and rotation), and even worsen adversarial robustness if used in isolation; and (ii) adversarial robustness is significantly affected by the augmentation probability, conversely to what is claimed in recent work. We conclude by discussing how to rethink the development and evaluation of novel data augmentation methods for adversarial robustness. Our open-source code is available at https://github.com/eghbalz/rethink_da_for_ar.
科研通智能强力驱动
Strongly Powered by AbleSci AI