Papers
Topics
Authors
Recent
Search
2000 character limit reached

Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training

Published 5 Dec 2017 in cs.CV, cs.DC, cs.LG, and stat.ML | (1712.01887v3)

Abstract: Large-scale distributed training requires significant communication bandwidth for gradient exchange that limits the scalability of multi-node training, and requires expensive high-bandwidth network infrastructure. The situation gets even worse with distributed training on mobile devices (federated learning), which suffers from higher latency, lower throughput, and intermittent poor connections. In this paper, we find 99.9% of the gradient exchange in distributed SGD is redundant, and propose Deep Gradient Compression (DGC) to greatly reduce the communication bandwidth. To preserve accuracy during compression, DGC employs four methods: momentum correction, local gradient clipping, momentum factor masking, and warm-up training. We have applied Deep Gradient Compression to image classification, speech recognition, and language modeling with multiple datasets including Cifar10, ImageNet, Penn Treebank, and Librispeech Corpus. On these scenarios, Deep Gradient Compression achieves a gradient compression ratio from 270x to 600x without losing accuracy, cutting the gradient size of ResNet-50 from 97MB to 0.35MB, and for DeepSpeech from 488MB to 0.74MB. Deep gradient compression enables large-scale distributed training on inexpensive commodity 1Gbps Ethernet and facilitates distributed training on mobile. Code is available at: https://github.com/synxlin/deep-gradient-compression.

Citations (1,323)

Summary

  • The paper introduces Deep Gradient Compression, a technique that reduces communication bandwidth by up to 600x through methods like gradient sparsification and momentum correction while preserving accuracy.
  • The methodology integrates gradient sparsification, momentum correction, local gradient clipping, momentum factor masking, and warm-up training to optimize distributed training.
  • Experimental results confirm improved scalability and efficiency, enabling robust training across various architectures and low-bandwidth networks, including mobile devices.

Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training

In the paper "Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training," the authors address a critical bottleneck in large-scale distributed training: the significant communication bandwidth required for gradient exchange. Synchronous SGD, a widely used method for distributed training, often incurs high communication costs that can dwarf the computational savings obtained by parallelizing processing across multiple nodes. This issue is exacerbated in federated learning scenarios, where mobile devices with unreliable and slow network connections participate in distributed training.

Key Techniques

The authors introduce Deep Gradient Compression (DGC), a framework that drastically reduces communication bandwidth without sacrificing model accuracy. DGC employs several key techniques:

  1. Gradient Sparsification: Only significant gradients (ones with magnitudes above a threshold) are transmitted, reducing the volume of data exchanged.
  2. Momentum Correction: Adjusts for the discounting factor between sparse updates to ensure convergence properties analogous to dense updates.
  3. Local Gradient Clipping: Prevents the explosion of gradients by clipping them before local accumulation.
  4. Momentum Factor Masking: Prevents stale momentum by masking momentum updates for delayed gradients.
  5. Warm-up Training: Gradually increases gradient sparsity at early training stages to adaptively handle the rapid changes in network parameters.

Experimental Verification

The authors validate DGC across several neural network architectures (CNNs and RNNs) and tasks:

  • Image Classification: Using ResNet-110 on Cifar10 and ResNet-50 on ImageNet.
  • Language Modeling: Using a 2-layer LSTM on the Penn Treebank dataset.
  • Speech Recognition: Using DeepSpeech on AN4 and Librispeech corpora.

The experiments demonstrate that DGC achieves gradient compression ratios ranging from 270x to 600x without compromising on model accuracy. For instance, DGC was able to reduce the gradient size of ResNet-50 from 97MB to 0.35MB and DeepSpeech from 488MB to 0.74MB. The detailed results of the experiments are summarized as follows:

Image Classification

  • Cifar10: DGC preserved or even slightly improved the accuracy across different batch sizes and scales.
  • ImageNet: DGC achieves 597x compression on AlexNet and 277x on ResNet-50 while maintaining the training accuracy close to the baseline.

Language Modeling

  • Penn Treebank: DGC maintained the perplexity close to the baseline while achieving a 462x compression rate.

Speech Recognition

  • LibriSpeech: DGC showed slight improvements in word error rate on both clean and noisy speech datasets while achieving 608x compression.

Implications and Future Directions

The results suggest significant implications for both practical and theoretical aspects of distributed training:

  1. Scalability: By reducing the required communication bandwidth, DGC enables training across a larger number of nodes even on inexpensive, low-bandwidth networks, such as 1Gbps Ethernet.
  2. Accessibility: DGC facilitates federated learning on mobile devices by mitigating the adverse effects of limited and intermittent network connectivity.
  3. Efficiency: DGC considerably reduces the operational costs associated with bandwidth and improves the speedup and scalability of distributed training frameworks.
  4. Robustness: By preserving training accuracy across a wide range of compression ratios and architectures, DGC establishes a robust groundwork for further research into efficient communication strategies in distributed machine learning systems.

Conclusion

The paper puts forth Deep Gradient Compression as an efficient, scalable method for reducing the communication bandwidth in distributed training without accuracy loss. The extensive empirical validation across various neural network models and datasets demonstrates that DGC can achieve compression ratios of up to 600x, significantly alleviating the communication bottleneck. Future research may investigate applying DGC broadly across diverse neural architectures and further optimize its implementation for specific hardware and network configurations.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 5 tweets with 34 likes about this paper.