后门
计算机科学
水准点(测量)
编码(集合论)
训练集
深层神经网络
人工智能
特洛伊木马
机器学习
人工神经网络
深度学习
计算机安全
大地测量学
集合(抽象数据类型)
程序设计语言
地理
作者
Kuofeng Gao,Yang Bai,Jindong Gu,Yong Yang,Shu‐Tao Xia
标识
DOI:10.1109/cvpr52729.2023.00390
摘要
Backdoor defenses have been studied to alleviate the threat of deep neural networks (DNNs) being backdoor attacked and thus maliciously altered. Since DNNs usually adopt some external training data from an untrusted third party, a robust backdoor defense strategy during the training stage is of importance. We argue that the core of training-time defense is to select poisoned samples and to handle them properly. In this work, we summarize the training-time defenses from a unified framework as splitting the poisoned dataset into two data pools. Under our framework, we propose an adaptively splitting dataset-based defense (ASD). Concretely, we apply loss-guided split and meta-learning-inspired split to dynamically update two data pools. With the split clean data pool and polluted data pool, ASD successfully defends against backdoor attacks during training. Extensive experiments on multiple benchmark datasets and DNN models against six state-of-the-art backdoor attacks demonstrate the superiority of our ASD. Our code is available at https://github.com/KuofengGao/ASD.
科研通智能强力驱动
Strongly Powered by AbleSci AI