计算机科学
加速
支持向量机
趋同(经济学)
同步(交流)
利用
集合(抽象数据类型)
机器学习
人工智能
解析
并行计算
算法
经济增长
计算机安全
频道(广播)
经济
程序设计语言
计算机网络
作者
Jiantong Jiang,Zeyi Wen,Zeke Wang,Bingsheng He,Jian Chen
标识
DOI:10.1109/tpds.2021.3101155
摘要
Structured Support Vector Machines (structured SVMs) are a fundamental machine learning algorithm, and have solid theoretical foundation and high effectiveness in applications such as natural language parsing and computer vision. However, training structured SVMs is very time-consuming, due to the large number of constraints and inferior convergence rates, especially for large training data sets. The high cost of training structured SVMs has hindered its adoption to new applications. In this article, we aim to improve the efficiency of structured SVMs by proposing a parallel and distributed solution (namely FastSSVM ) for training structured SVMs building on top of MPI and OpenMP. FastSSVM exploits a series of optimizations (e.g., optimizations on data storage and synchronization) to efficiently use the resources of the nodes in a cluster and the cores of the nodes. Moreover, FastSSVM tackles the large constraint set problem by batch processing and addresses the slow convergence challenge by adapting stop conditions based on the improvement of each iteration. We theoretically prove that our solution is guaranteed to converge to a global optimum. A comprehensive experimental study shows that FastSSVM can achieve at least four times speedup over the existing solutions, and in some cases can achieve two to three orders of magnitude speedup.
科研通智能强力驱动
Strongly Powered by AbleSci AI