基于Bagging-Down SGD算法的分布式深度网络

Translated title of the contribution: Distributed deep networks based on Bagging-Down SGD algorithm

Chao Qin, Xiaoguang Gao, Daqing Chen

Research output: Contribution to journalArticlepeer-review

2 Scopus citations

Abstract

As a cutting-edge disruptive technology, deep learning and unsupervised learning have attracted a significant research attention, and it has been widely acknowledged that training big data with a distributed deep learning algorithm can get better structures. However, there are two main problems with traditional distributed deep learning algorithms: the speed of training is slow and the accuracy of training is low. The Bootstrap aggregating-down stochastic gradient descent (Bagging-Down SGD) algorithm is proposed to solve the speed problem mainly. We add a speed controller to update the parameters of the single machine statistically, and to split model training and parameters updating to improve the training speed with the assurance of the same accuracy. It is to be proved in the experiment that the algorithm has the generality to learn the structures of different kinds of data.

Translated title of the contributionDistributed deep networks based on Bagging-Down SGD algorithm
Original languageChinese (Traditional)
Pages (from-to)1021-1027
Number of pages7
JournalXi Tong Gong Cheng Yu Dian Zi Ji Shu/Systems Engineering and Electronics
Volume41
Issue number5
DOIs
StatePublished - 1 May 2019

Fingerprint

Dive into the research topics of 'Distributed deep networks based on Bagging-Down SGD algorithm'. Together they form a unique fingerprint.

Cite this