File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
  • Find via Find It@HKUL
Supplementary

Conference Paper: Differentiable Dynamic Quantization with Mixed Precision and Adaptive Resolution

TitleDifferentiable Dynamic Quantization with Mixed Precision and Adaptive Resolution
Authors
Issue Date2021
PublisherML Research Press. The Journal's web site is located at http://proceedings.mlr.press/
Citation
The 38th International Conference on Machine Learning (ICML), Virtual Conference, 18-24 July 2021. In Proceedings of Machine Learning Research (PMLR), v. 139: Proceedings of ICML 2021, p. 12546-12556 How to Cite?
AbstractModel quantization is challenging due to many tedious hyper-parameters such as precision (bitwidth), dynamic range (minimum and maximum discrete values) and stepsize (interval between discrete values). Unlike prior arts that carefully tune these values, we present a fully differentiable approach to learn all of them, named Differentiable Dynamic Quantization (DDQ), which has several benefits. (1) DDQ is able to quantize challenging lightweight architectures like MobileNets, where different layers prefer different quantization parameters. (2) DDQ is hardware-friendly and can be easily implemented using low-precision matrix-vector multiplication, making it capable in many hardware such as ARM. (3) Extensive experiments show that DDQ outperforms prior arts on many networks and benchmarks, especially when models are already efficient and compact. e.g., DDQ is the first approach that achieves lossless 4-bit quantization for MobileNetV2 on ImageNet.
DescriptionApplications (CV and NLP) Session
Persistent Identifierhttp://hdl.handle.net/10722/301433
ISSN

 

DC FieldValueLanguage
dc.contributor.authorZhang, Z-
dc.contributor.authorShao, W-
dc.contributor.authorGu, J-
dc.contributor.authorWang, X-
dc.contributor.authorLuo, P-
dc.date.accessioned2021-07-27T08:11:00Z-
dc.date.available2021-07-27T08:11:00Z-
dc.date.issued2021-
dc.identifier.citationThe 38th International Conference on Machine Learning (ICML), Virtual Conference, 18-24 July 2021. In Proceedings of Machine Learning Research (PMLR), v. 139: Proceedings of ICML 2021, p. 12546-12556-
dc.identifier.issn2640-3498-
dc.identifier.urihttp://hdl.handle.net/10722/301433-
dc.descriptionApplications (CV and NLP) Session-
dc.description.abstractModel quantization is challenging due to many tedious hyper-parameters such as precision (bitwidth), dynamic range (minimum and maximum discrete values) and stepsize (interval between discrete values). Unlike prior arts that carefully tune these values, we present a fully differentiable approach to learn all of them, named Differentiable Dynamic Quantization (DDQ), which has several benefits. (1) DDQ is able to quantize challenging lightweight architectures like MobileNets, where different layers prefer different quantization parameters. (2) DDQ is hardware-friendly and can be easily implemented using low-precision matrix-vector multiplication, making it capable in many hardware such as ARM. (3) Extensive experiments show that DDQ outperforms prior arts on many networks and benchmarks, especially when models are already efficient and compact. e.g., DDQ is the first approach that achieves lossless 4-bit quantization for MobileNetV2 on ImageNet.-
dc.languageeng-
dc.publisherML Research Press. The Journal's web site is located at http://proceedings.mlr.press/-
dc.relation.ispartofProceedings of Machine Learning Research (PMLR)-
dc.relation.ispartofThe 38th International Conference on Machine Learning (ICML), 2021-
dc.titleDifferentiable Dynamic Quantization with Mixed Precision and Adaptive Resolution-
dc.typeConference_Paper-
dc.identifier.emailLuo, P: pluo@hku.hk-
dc.identifier.authorityLuo, P=rp02575-
dc.identifier.hkuros323759-
dc.identifier.volume139: Proceedings of ICML 2021-
dc.identifier.spage12546-
dc.identifier.epage12556-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats