Communication Usage Optimization of Gradient Sparsification with Aggregation
碩士 === 國立臺灣大學 === 資訊工程學研究所 === 106 === Communication usage is a bottleneck of scaling workers for distributed deep learning. One solution is to compress the exchanged gradients into sparse format with gradient sparsification. We found that the send cost of server, which is the aggregated size of spa...
Main Authors: | Sheng-Ping Wang, 王盛平 |
---|---|
Other Authors: | Pangfeng Liu |
Format: | Others |
Language: | en_US |
Published: |
2018
|
Online Access: | http://ndltd.ncl.edu.tw/handle/ppyyqb |
Similar Items
-
Spectral sparsification
by: Camacho, Martin Ayalde
Published: (2014) -
On the combinatorics of sparsification
by: Huang Fenix WD, et al.
Published: (2012-10-01) -
Automatic DenseNet Sparsification
by: Tao Li, et al.
Published: (2020-01-01) -
Compilation by stochastic Hamiltonian sparsification
by: Yingkai Ouyang, et al.
Published: (2020-02-01) -
Vertex Sparsification and Oblivious Reductions
by: Moitra, Ankur
Published: (2014)