对抗制
强化学习
计算机科学
人工智能
深度学习
分类器(UML)
机器学习
脆弱性(计算)
计算机安全
作者
R Praveen Kumar,I Niranjan Kumar,Sujith Sivasankaran,A. Mohan Vamsi,Vineeth Vijayaraghavan
标识
DOI:10.1109/icmla52953.2021.00279
摘要
Deep learning plays a vital role in day-to-day applications. Recent studies show that deep learning models are not resilient against adversarial attacks, which is also applicable to Deep Reinforcement Learning (DRL) agents. Considering sensitive use cases of the DRL agents, there is a pressing need to make them robust to adversarial attacks. However, to design an efficient defense, it is imperative that we fully understand the vulnerability of such agents. In this work, we propose statistical and model-based approaches to identify critical states in an episode. Our work shows that by attacking less than 1% of the total number of states, the agent performance can be reduced by more than 40%. Furthermore, we model a long-term impact classifier to identify critical states. This method reduces the average compute time by 80.3% when compared to previous approaches.
科研通智能强力驱动
Strongly Powered by AbleSci AI