反向传播
趋同(经济学)
人工神经网络
随机微分方程
离散化
计算机科学
数学
解算器
数学优化
MNIST数据库
算法
水准点(测量)
应用数学
人工智能
经济
经济增长
数学分析
大地测量学
地理
作者
Richard Archibald,Feng Bao,Yanzhao Cao,Hui Sun
摘要
.The aim of this paper is to carry out convergence analysis and algorithm implementation of a novel sample-wise backpropagation method for training a class of stochastic neural networks (SNNs). The preliminary discussion on such an SNN framework was first introduced in [Archibald et al., Discrete Contin. Dyn. Syst. Ser. S, 15 (2022), pp. 2807–2835]. The structure of the SNN is formulated as a discretization of a stochastic differential equation (SDE). A stochastic optimal control framework is introduced to model the training procedure, and a sample-wise approximation scheme for the adjoint backward SDE is applied to improve the efficiency of the stochastic optimal control solver, which is equivalent to the backpropagation for training the SNN. The convergence analysis is derived by introducing a novel joint conditional expectation for the gradient process. Under the convexity assumption, our result indicates that the number of SNN training steps should be proportional to the square of the number of layers in the convex optimization case. In the implementation of the sample-based SNN algorithm with the benchmark MNIST dataset, we adopt the convolution neural network (CNN) architecture and demonstrate that our sample-based SNN algorithm is more robust than the conventional CNN.Keywordsprobabilistic learningstochastic neural networksconvergence analysisbackward stochastic differential equationsstochastic gradient descentMSC codes65C2060H1060H30
科研通智能强力驱动
Strongly Powered by AbleSci AI