线性子空间
数学
黑森矩阵
Krylov子空间
曲率
牛顿法
订单(交换)
应用数学
子空间拓扑
班级(哲学)
空格(标点符号)
组合数学
纯数学
数学分析
线性系统
非线性系统
几何学
计算机科学
人工智能
经济
物理
操作系统
量子力学
财务
作者
Serge Gratton,Sadok Jerad,Philippe L. Toint
标识
DOI:10.1093/imanum/drae021
摘要
Abstract A class of second-order algorithms is proposed for minimizing smooth nonconvex functions that alternates between regularized Newton and negative curvature steps in an iteration-dependent subspace. In most cases, the Hessian matrix is regularized with the square root of the current gradient and an additional term taking moderate negative curvature into account, a negative curvature step being taken only exceptionally. Practical variants are detailed where the subspaces are chosen to be the full space, or Krylov subspaces. In the first case, the proposed method only requires the solution of a single linear system at nearly all iterations. We establish that at most $\mathcal{O}\big ( |\!\log \epsilon |\,\epsilon ^{-3/2}\big )$ evaluations of the problem’s objective function and derivatives are needed for algorithms in the new class to obtain an $\epsilon $-approximate first-order minimizer, and at most $\mathcal{O}\big (|\!\log \epsilon |\,\epsilon ^{-3}\big )$ to obtain a second-order one. Encouraging initial numerical experiments with two full-space and two Krylov-subspaces variants are finally presented.
科研通智能强力驱动
Strongly Powered by AbleSci AI