过度拟合
计算机科学
残余物
人工智能
正规化(语言学)
机器学习
分位数
稳健性(进化)
基本事实
深度学习
数学
算法
计量经济学
人工神经网络
生物化学
基因
化学
作者
Hong Liu,Zhun Zhong,Nicu Sebe,Shin’ichi Satoh
标识
DOI:10.1016/j.artint.2023.103877
摘要
Overfitting in adversarial training has attracted the interest of researchers in the community of artificial intelligence and machine learning in recent years. To address this issue, in this paper we begin by evaluating the defense performances of several calibration methods on various robust models. Our analysis and experiments reveal two intriguing properties: 1) a well-calibrated robust model is decreasing the confidence of robust model; 2) there is a trade-off between the confidences of natural and adversarial images. These new properties offer a straightforward insight into designing a simple but effective regularization, called Self-Residual-Calibration (SRC). The proposed SRC calculates the absolute residual between adversarial and natural logit features corresponding to the ground-truth labels. Furthermore, we utilize the pinball loss to minimize the quantile residual between them, resulting in more robust regularization. Extensive experiments indicate that our SRC can effectively mitigate the overfitting problem while improving the robustness of state-of-the-art models. Importantly, SRC is complementary to various regularization methods. When combined with them, we are capable of achieving the top-rank performance on the AutoAttack benchmark leaderboard.
科研通智能强力驱动
Strongly Powered by AbleSci AI