数学
下降(航空)
梯度下降
应用数学
随机梯度下降算法
数学优化
计算机科学
人工智能
人工神经网络
工程类
航空航天工程
作者
Wenqing Ouyang,Yang Liu,Andre Milzarek
出处
期刊:Siam Journal on Optimization
[Society for Industrial and Applied Mathematics]
日期:2024-01-19
卷期号:34 (1): 336-365
摘要
.Anderson acceleration (\(\mathsf{AA}\)) is a popular acceleration technique to enhance the convergence of fixed-point schemes. The analysis of \(\mathsf{AA}\) approaches often focuses on the convergence behavior of a corresponding fixed-point residual, while the behavior of the underlying objective function values along the accelerated iterates is currently not well understood. In this paper, we investigate local properties of \(\mathsf{AA}\) with restarting applied to a basic gradient scheme (\(\mathsf{AA}\mbox{-}\mathsf{R}\)) in terms of function values. Specifically, we show that \(\mathsf{AA}\mbox{-}\mathsf{R}\) is a local descent method and that it can decrease the objective function at a rate no slower than the gradient method up to higher-order error terms. These new results theoretically support the good numerical performance of \(\mathsf{AA}(\mbox{-}\mathsf{R})\) when heuristic descent conditions are used for globalization and they provide a novel perspective on the convergence analysis of \(\mathsf{AA}\mbox{-}\mathsf{R}\) that is more amenable to nonconvex optimization problems. Numerical experiments are conducted to illustrate our theoretical findings.KeywordsAnderson accelerationdescent propertiesrestartingMSC codes90C3065K0590C0690C53
科研通智能强力驱动
Strongly Powered by AbleSci AI