波束搜索
计算机科学
排名(信息检索)
人工智能
词(群论)
机器学习
贪婪算法
对抗制
分布语义学
空格(标点符号)
搜索算法
算法
数学
语义相似性
几何学
操作系统
作者
Bin Zhu,Zhaoquan Gu,Yaguan Qian,Francis C. M. Lau,Zhihong Tian
出处
期刊:Neurocomputing
[Elsevier BV]
日期:2022-08-01
卷期号:500: 135-142
被引量:8
标识
DOI:10.1016/j.neucom.2022.05.054
摘要
Adversarial attacks in NLP are difficult to ward off because of the discrete and highly abstract nature of human languages. Prior works utilize different word replacement strategies to generate semantic-preserving adversarial texts. These query-based methods, however, have limited exploration of the search space. To fully explore the search space, an improved beam search with multiple random perturbing positions is used. Besides, we use the transferable vulnerability from surrogate models to choose vulnerable candidate words for target models. We empirically show that beam search with multiple random attacking positions works better than the commonly used greedy search with word importance ranking. Extensive experiments on three popular datasets demonstrate that our method can outperform three advanced attacking methods under black-box settings. We provide ablation studies to clearly show the effectiveness of our improved beam search which can achieve a higher success rate than the greedy approach under the same query budget.
科研通智能强力驱动
Strongly Powered by AbleSci AI