过度拟合
后验概率
计算机科学
迭代函数
贝叶斯概率
蒙特卡罗方法
吉布斯抽样
人工智能
趋同(经济学)
采样(信号处理)
算法
数学
统计
人工神经网络
滤波器(信号处理)
数学分析
经济增长
计算机视觉
经济
作者
Max Welling,Yee Whye Teh
摘要
In this paper we propose a new framework for learning from large scale datasets based on iterative learning from small mini-batches. By adding the right amount of noise to a standard stochastic gradient optimization algorithm we show that the iterates will converge to samples from the true posterior distribution as we anneal the stepsize. This seamless transition between optimization and Bayesian posterior provides an inbuilt protection against overfitting. We also propose a practical method for Monte Carlo estimates of posterior statistics which monitors a sampling threshold and collects samples after it has been surpassed. We apply the method to three models: a mixture of Gaussians, logistic regression and ICA with natural gradients.
科研通智能强力驱动
Strongly Powered by AbleSci AI