随机梯度下降算法
交叉口(航空)
趋同(经济学)
凸函数
数学优化
正多边形
数学
收敛速度
计算机科学
可分离空间
功能(生物学)
约束(计算机辅助设计)
算法
人工智能
生物
进化生物学
频道(广播)
工程类
数学分析
人工神经网络
航空航天工程
经济
经济增长
计算机网络
几何学
作者
Thinh T. Doan,Joseph Lubars,Carolyn L. Beck,R. Srikant
标识
DOI:10.1016/j.ifacol.2018.12.064
摘要
Stochastic gradient descent is a common algorithm used in machine learning, especially when the loss function is in a separable form. Here, we consider SGD for constrained convex optimization problems, where the constraint is expressed as an intersection of a large number of convex sets. It is expensive to project the result of the gradient descent algorithm on each of these convex sets at each iteration; thus, there is a growing body of work which considers projections onto a random subset of the convex sets. In this paper, we consider a distributed version (parameter-server model) of the random projections since a centralized approach is too slow for very large-scale problems. Our main result is as follows: under a mild regularity condition on the convex sets, we show that the rate of convergence of distributed SGD with distributed random projections is the same as that of distributed SGD applied to a problem with no constraints, except for a factor which captures the regularity assumption.
科研通智能强力驱动
Strongly Powered by AbleSci AI