File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/BigData.2018.8622264
- Scopus: eid_2-s2.0-85062613824
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Conference Paper: Scheduling Large-scale Distributed Training via Reinforcement Learning
Title | Scheduling Large-scale Distributed Training via Reinforcement Learning |
---|---|
Authors | |
Keywords | Convolutional Neural Network Reinforcement Learning Optimization Deep Learning |
Issue Date | 2019 |
Citation | Proceedings - 2018 IEEE International Conference on Big Data, Big Data 2018, 2019, p. 1797-1806 How to Cite? |
Abstract | © 2018 IEEE. Scheduling the training procedure of deep neural networks (DNNs) such as tuning the learning rates is crucial to the successes of deep learning. Previous strategies such as piecewise and exponential learning rate schedulers have different arguments (hyper-parameters) that need to be tuned manually. With the expanding of data scale and model computation, searching for these arguments spends lots of empirical efforts. To address this issue, this work proposes policy schedular that determines the arguments of learning rate (lr) by reinforcement learning, significantly reducing costs to tune them. The policy schedular possesses several appealing benefits. First, instead of manually defining the values of initial lr and ultimate lr, it autonomously determines these values in training. Second, rather than using predefined functions to update lr, it adaptively oscillates lr by monitoring learning curves without human intervention. Third, it is able to select lr for each block or layer of a DNN. Experiments show that the DNNs trained with policy schedular achieve superior performances, outperforming previous work on various tasks and benchmarks such as ImageNet, COCO, and learning-to-learn. |
Persistent Identifier | http://hdl.handle.net/10722/273691 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Peng, Zhanglin | - |
dc.contributor.author | Ren, Jiamin | - |
dc.contributor.author | Zhang, Ruimao | - |
dc.contributor.author | Wu, Lingyun | - |
dc.contributor.author | Wang, Xinjiang | - |
dc.contributor.author | Luo, Ping | - |
dc.date.accessioned | 2019-08-12T09:56:22Z | - |
dc.date.available | 2019-08-12T09:56:22Z | - |
dc.date.issued | 2019 | - |
dc.identifier.citation | Proceedings - 2018 IEEE International Conference on Big Data, Big Data 2018, 2019, p. 1797-1806 | - |
dc.identifier.uri | http://hdl.handle.net/10722/273691 | - |
dc.description.abstract | © 2018 IEEE. Scheduling the training procedure of deep neural networks (DNNs) such as tuning the learning rates is crucial to the successes of deep learning. Previous strategies such as piecewise and exponential learning rate schedulers have different arguments (hyper-parameters) that need to be tuned manually. With the expanding of data scale and model computation, searching for these arguments spends lots of empirical efforts. To address this issue, this work proposes policy schedular that determines the arguments of learning rate (lr) by reinforcement learning, significantly reducing costs to tune them. The policy schedular possesses several appealing benefits. First, instead of manually defining the values of initial lr and ultimate lr, it autonomously determines these values in training. Second, rather than using predefined functions to update lr, it adaptively oscillates lr by monitoring learning curves without human intervention. Third, it is able to select lr for each block or layer of a DNN. Experiments show that the DNNs trained with policy schedular achieve superior performances, outperforming previous work on various tasks and benchmarks such as ImageNet, COCO, and learning-to-learn. | - |
dc.language | eng | - |
dc.relation.ispartof | Proceedings - 2018 IEEE International Conference on Big Data, Big Data 2018 | - |
dc.subject | Convolutional Neural Network | - |
dc.subject | Reinforcement Learning | - |
dc.subject | Optimization | - |
dc.subject | Deep Learning | - |
dc.title | Scheduling Large-scale Distributed Training via Reinforcement Learning | - |
dc.type | Conference_Paper | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1109/BigData.2018.8622264 | - |
dc.identifier.scopus | eid_2-s2.0-85062613824 | - |
dc.identifier.spage | 1797 | - |
dc.identifier.epage | 1806 | - |