对抗制
显著性(神经科学)
计算机科学
关系(数据库)
人工智能
模式识别(心理学)
数据挖掘
作者
Luoqiu Li,Xiang Chen,Zhen Bi,Xin Xie,Shumin Deng,Ningyu Zhang,Chuanqi Tan,Mosha Chen,Huajun Chen
标识
DOI:10.1145/3502223.3502237
摘要
Recent neural-based relation extraction approaches, though achieving promising improvement on benchmark datasets, have reported their vulnerability towards adversarial attacks. Thus far, efforts mostly focused on generating adversarial samples or defending adversarial attacks, but little is known about the difference between normal and adversarial samples. In this work, we take the first step to leverage the salience-based method to analyze those adversarial samples. We observe that salience tokens have a direct correlation with adversarial perturbations. We further find the adversarial perturbations are either those tokens not existing in the training set or superficial cues associated with relation labels. To some extent, our approach unveils the characters against adversarial samples. We release an open-source testbed, "DiagnoseAdv"1, for future research purposes.
科研通智能强力驱动
Strongly Powered by AbleSci AI