File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Characterizing Deep Learning Training Workloads on Alibaba-PAI

TitleCharacterizing Deep Learning Training Workloads on Alibaba-PAI
Authors
Issue Date2019
PublisherIEEE. The Journal's web site is located at https://ieeexplore.ieee.org/xpl/conhome/1000819/all-proceedings
Citation
Proceedings of 2019 IEEE International Symposium on Workload Characterization (IISWC), Orlando, FL, USA, 3-5 November 2019, p. 189-202 How to Cite?
AbstractModern deep learning models have been exploited in various domains, including computer vision (CV), natural language processing (NLP), search and recommendation. In practical AI clusters, workloads training these models are run using software frameworks such as TensorFlow, Caffe, PyTorch and CNTK. One critical issue for efficiently operating practical AI clouds, is to characterize the computing and data transfer demands of these workloads, and more importantly, the training performance given the underlying software framework and hardware configurations. In this paper, we characterize deep learning training workloads from Platform of Artificial Intelligence (PAI) in Alibaba. We establish an analytical framework to investigate detailed execution time breakdown of various workloads using different training architectures, to identify performance bottleneck. Results show that weight/gradient communication during training takes almost 62% of the total execution time among all our workloads on average. The computation part, involving both GPU computing and memory access, are not the biggest bottleneck based on collective behavior of the workloads. We further evaluate attainable performance of the workloads on various potential software/hardware mappings, and explore implications on software architecture selection and hardware configurations. We identify that 60% of PS/Worker workloads can be potentially sped up when ported to the AllReduce architecture exploiting the high-speed NVLink for GPU interconnect, and on average 1.7X speedup can be achieved when Ethernet bandwidth is upgraded from 25 Gbps to 100 Gbps.
Persistent Identifierhttp://hdl.handle.net/10722/301419
ISBN

 

DC FieldValueLanguage
dc.contributor.authorWang, M-
dc.contributor.authorMeng, C-
dc.contributor.authorLong, G-
dc.contributor.authorWu, C-
dc.contributor.authorYang, J-
dc.contributor.authorLin, W-
dc.contributor.authorJia, Y-
dc.date.accessioned2021-07-27T08:10:47Z-
dc.date.available2021-07-27T08:10:47Z-
dc.date.issued2019-
dc.identifier.citationProceedings of 2019 IEEE International Symposium on Workload Characterization (IISWC), Orlando, FL, USA, 3-5 November 2019, p. 189-202-
dc.identifier.isbn9781728140469-
dc.identifier.urihttp://hdl.handle.net/10722/301419-
dc.description.abstractModern deep learning models have been exploited in various domains, including computer vision (CV), natural language processing (NLP), search and recommendation. In practical AI clusters, workloads training these models are run using software frameworks such as TensorFlow, Caffe, PyTorch and CNTK. One critical issue for efficiently operating practical AI clouds, is to characterize the computing and data transfer demands of these workloads, and more importantly, the training performance given the underlying software framework and hardware configurations. In this paper, we characterize deep learning training workloads from Platform of Artificial Intelligence (PAI) in Alibaba. We establish an analytical framework to investigate detailed execution time breakdown of various workloads using different training architectures, to identify performance bottleneck. Results show that weight/gradient communication during training takes almost 62% of the total execution time among all our workloads on average. The computation part, involving both GPU computing and memory access, are not the biggest bottleneck based on collective behavior of the workloads. We further evaluate attainable performance of the workloads on various potential software/hardware mappings, and explore implications on software architecture selection and hardware configurations. We identify that 60% of PS/Worker workloads can be potentially sped up when ported to the AllReduce architecture exploiting the high-speed NVLink for GPU interconnect, and on average 1.7X speedup can be achieved when Ethernet bandwidth is upgraded from 25 Gbps to 100 Gbps.-
dc.languageeng-
dc.publisherIEEE. The Journal's web site is located at https://ieeexplore.ieee.org/xpl/conhome/1000819/all-proceedings-
dc.relation.ispartofIEEE International Symposium on Workload Characterization (IISWC)-
dc.rightsIEEE International Symposium on Workload Characterization (IISWC). Copyright © IEEE.-
dc.rights©2019 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.-
dc.titleCharacterizing Deep Learning Training Workloads on Alibaba-PAI-
dc.typeConference_Paper-
dc.identifier.emailWu, C: cwu@cs.hku.hk-
dc.identifier.authorityWu, C=rp01397-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/IISWC47752.2019.9042047-
dc.identifier.scopuseid_2-s2.0-85083107261-
dc.identifier.hkuros323518-
dc.identifier.spage189-
dc.identifier.epage202-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats