File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization Framework

TitleMix and Match: A Novel FPGA-Centric Deep Neural Network Quantization Framework
Authors
KeywordsEmbedded systems
Field programmable gate arrays
Learning (artificial intelligence)
Matrix multiplication
Neural nets
Issue Date2021
PublisherIEEE Computer Society.
Citation
IEEE International Symposium on High-Performance Computer Architecture (HPCA) (Virtual Event), Seoul, Korea, 27 February-3 March, 2021. In Proceedings: 27th IEEE International Symposium on High-Performance Computer Architecture, 27 February-3 March 2021, p. 208-220 How to Cite?
AbstractDeep Neural Networks (DNNs) have achieved extraordinary performance in various application domains. To support diverse DNN models, efficient implementations of DNN inference on edge-computing platforms, e.g., ASICs, FPGAs, and embedded systems, are extensively investigated. Due to the huge model size and computation amount, model compression is a critical step to deploy DNN models on edge devices. This paper focuses on weight quantization, a hardware-friendly model compression approach that is complementary to weight pruning. Unlike existing methods that use the same quantization scheme for all weights, we propose the first solution that applies different quantization schemes for different rows of the weight matrix. It is motivated by (1) the distribution of the weights in the different rows are not the same; and (2) the potential of achieving better utilization of heterogeneous FPGA hardware resources. To achieve that, we first propose a hardware-friendly quantization scheme named sum-of-power-of-2 (SP2) suitable for Gaussian-like weight distribution, in which the multiplication arithmetic can be replaced with logic shifter and adder, thereby enabling highly efficient implementations with the FPGA LUT resources. In contrast, the existing fixed-point quantization is suitable for Uniform-like weight distribution and can be implemented efficiently by DSP. Then to fully explore the resources, we propose an FPGA-centric mixed scheme quantization (MSQ) with an ensemble of the proposed SP2 and the fixed-point schemes. Combining the two schemes can maintain, or even increase accuracy due to better matching with weight distributions.
Persistent Identifierhttp://hdl.handle.net/10722/323349
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorChang, SE-
dc.contributor.authorLi, Y-
dc.contributor.authorSun, M-
dc.contributor.authorShi, R-
dc.contributor.authorSo, HKH-
dc.contributor.authorQian, X-
dc.contributor.authorWang, Y-
dc.contributor.authorLin, X-
dc.date.accessioned2022-12-09T10:45:18Z-
dc.date.available2022-12-09T10:45:18Z-
dc.date.issued2021-
dc.identifier.citationIEEE International Symposium on High-Performance Computer Architecture (HPCA) (Virtual Event), Seoul, Korea, 27 February-3 March, 2021. In Proceedings: 27th IEEE International Symposium on High-Performance Computer Architecture, 27 February-3 March 2021, p. 208-220-
dc.identifier.urihttp://hdl.handle.net/10722/323349-
dc.description.abstractDeep Neural Networks (DNNs) have achieved extraordinary performance in various application domains. To support diverse DNN models, efficient implementations of DNN inference on edge-computing platforms, e.g., ASICs, FPGAs, and embedded systems, are extensively investigated. Due to the huge model size and computation amount, model compression is a critical step to deploy DNN models on edge devices. This paper focuses on weight quantization, a hardware-friendly model compression approach that is complementary to weight pruning. Unlike existing methods that use the same quantization scheme for all weights, we propose the first solution that applies different quantization schemes for different rows of the weight matrix. It is motivated by (1) the distribution of the weights in the different rows are not the same; and (2) the potential of achieving better utilization of heterogeneous FPGA hardware resources. To achieve that, we first propose a hardware-friendly quantization scheme named sum-of-power-of-2 (SP2) suitable for Gaussian-like weight distribution, in which the multiplication arithmetic can be replaced with logic shifter and adder, thereby enabling highly efficient implementations with the FPGA LUT resources. In contrast, the existing fixed-point quantization is suitable for Uniform-like weight distribution and can be implemented efficiently by DSP. Then to fully explore the resources, we propose an FPGA-centric mixed scheme quantization (MSQ) with an ensemble of the proposed SP2 and the fixed-point schemes. Combining the two schemes can maintain, or even increase accuracy due to better matching with weight distributions.-
dc.languageeng-
dc.publisherIEEE Computer Society.-
dc.relation.ispartofProceedings: 27th IEEE International Symposium on High-Performance Computer Architecture, 27 February-3 March 2021-
dc.rightsProceedings: 27th IEEE International Symposium on High-Performance Computer Architecture, 27 February-3 March 2021. Copyright © IEEE Computer Society.-
dc.subjectEmbedded systems-
dc.subjectField programmable gate arrays-
dc.subjectLearning (artificial intelligence)-
dc.subjectMatrix multiplication-
dc.subjectNeural nets-
dc.titleMix and Match: A Novel FPGA-Centric Deep Neural Network Quantization Framework-
dc.typeConference_Paper-
dc.identifier.emailSo, HKH: hso@eee.hku.hk-
dc.identifier.authoritySo, HKH=rp00169-
dc.identifier.doi10.1109/HPCA51647.2021.00027-
dc.identifier.hkuros342981-
dc.identifier.spage208-
dc.identifier.epage220-
dc.identifier.isiWOS:000671076000016-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats