计算机科学
估计员
样品(材料)
人工神经网络
样本量测定
比例(比率)
随机性
蒙特卡罗方法
功能(生物学)
数学优化
人工智能
统计
数学
量子力学
色谱法
物理
化学
进化生物学
生物
作者
Kangning Wang,Benle Zhang,Xiaofei Sun,Shaomin Li
标识
DOI:10.1016/j.eswa.2022.116698
摘要
This paper proposes a communication-efficient pilot sample surrogate loss function framework that is able to solve efficient statistical estimation for a non-randomly distributed system. Specifically, we first collect a small size pilot sample from worker machines, then approximate the global loss function by a surrogate one, which only relates to the pilot sample and the gradients of local datasets. Without any restrictive condition about randomness, the established asymptotical properties show that the resulting estimator obtained by minimizing the surrogate loss is equivalent with the global estimator. Since the pilot sample and gradients can easily be communicated between the master and worker machines, the communication cost is significantly reduced. What is more, as a specific application, we apply our new method to the neural network with large-scale data for fast and accurate optimization. Monte Carlo simulations and real-world application are also used to validate our method. • Efficient estimation for a non-randomly distributed system is proposed. • The communication cost is significantly reduced. • Theoretical properties are established under mild conditions. • We apply the new method to the neural network with large-scale data.
科研通智能强力驱动
Strongly Powered by AbleSci AI