File Download

There are no files associated with this item.

Supplementary

Conference Paper: SAPipe: Staleness-Aware Pipeline for Data Parallel DNN Training

TitleSAPipe: Staleness-Aware Pipeline for Data Parallel DNN Training
Authors
Issue Date2022
PublisherCurran Associates, Inc.
Citation
Thirty-Sixth Conference on Neural Information Processing Systems, new Orleans, Louisiana, United States, November 28, 2022 How to Cite?
AbstractData parallelism across multiple machines is widely adopted for accelerating distributed deep learning, but it is hard to achieve linear speedup due to the heavy communication. In this paper, we propose SAPipe, a performant system that pushes the training speed of data parallelism to its fullest extent. By introducing partial staleness, the communication overlaps the computation with minimal staleness in SAPipe. To mitigate additional problems incurred by staleness, SAPipe adopts staleness compensation techniques including weight prediction and delay compensation with provably lower error bounds. Additionally, SAPipe presents an algorithm-system co-design with runtime optimization to minimize system overhead for the staleness training pipeline and staleness compensation. We have implemented SAPipe in the BytePS framework, compatible to both TensorFlow and PyTorch. Our experiments show that SAPipe achieves up to 157% speedups over BytePS (non-stale), and outperforms PipeSGD in accuracy by up to 13.7%.
Persistent Identifierhttp://hdl.handle.net/10722/320327

 

DC FieldValueLanguage
dc.contributor.authorChen, Y-
dc.contributor.authorXie, C-
dc.contributor.authorMa, M-
dc.contributor.authorGu, J-
dc.contributor.authorPeng, Y-
dc.contributor.authorLin, H-
dc.contributor.authorWu, C-
dc.contributor.authorZhu, Y-
dc.date.accessioned2022-10-21T07:51:14Z-
dc.date.available2022-10-21T07:51:14Z-
dc.date.issued2022-
dc.identifier.citationThirty-Sixth Conference on Neural Information Processing Systems, new Orleans, Louisiana, United States, November 28, 2022-
dc.identifier.urihttp://hdl.handle.net/10722/320327-
dc.description.abstractData parallelism across multiple machines is widely adopted for accelerating distributed deep learning, but it is hard to achieve linear speedup due to the heavy communication. In this paper, we propose SAPipe, a performant system that pushes the training speed of data parallelism to its fullest extent. By introducing partial staleness, the communication overlaps the computation with minimal staleness in SAPipe. To mitigate additional problems incurred by staleness, SAPipe adopts staleness compensation techniques including weight prediction and delay compensation with provably lower error bounds. Additionally, SAPipe presents an algorithm-system co-design with runtime optimization to minimize system overhead for the staleness training pipeline and staleness compensation. We have implemented SAPipe in the BytePS framework, compatible to both TensorFlow and PyTorch. Our experiments show that SAPipe achieves up to 157% speedups over BytePS (non-stale), and outperforms PipeSGD in accuracy by up to 13.7%.-
dc.languageeng-
dc.publisherCurran Associates, Inc.-
dc.titleSAPipe: Staleness-Aware Pipeline for Data Parallel DNN Training-
dc.typeConference_Paper-
dc.identifier.emailWu, C: cwu@cs.hku.hk-
dc.identifier.authorityWu, C=rp01397-
dc.identifier.hkuros340524-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats