File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: On the sparsity of neural machine translation models

TitleOn the sparsity of neural machine translation models
Authors
Issue Date2020
PublisherAssociation for Computational Linguistics.
Citation
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP 2020), Virtual Meeting, 16-20 November 2020, p. 1060–1066 How to Cite?
AbstractModern neural machine translation (NMT) models employ a large number of parameters, which leads to serious over-parameterization and typically causes the underutilization of computational resources. In response to this problem, we empirically investigate whether the redundant parameters can be reused to achieve better performance. Experiments and analyses are systematically conducted on different datasets and NMT architectures. We show that: 1) the pruned parameters can be rejuvenated to improve the baseline model by up to +0.8 BLEU points; 2) the rejuvenated parameters are reallocated to enhance the ability of modeling low-level lexical information.
DescriptionShort Paper - Gather Session 1A: Machine Translation and Multilinguality
Persistent Identifierhttp://hdl.handle.net/10722/287776

 

DC FieldValueLanguage
dc.contributor.authorWang, Y-
dc.contributor.authorWang, L-
dc.contributor.authorLi, VOK-
dc.contributor.authorTu, Z-
dc.date.accessioned2020-10-05T12:03:07Z-
dc.date.available2020-10-05T12:03:07Z-
dc.date.issued2020-
dc.identifier.citationProceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP 2020), Virtual Meeting, 16-20 November 2020, p. 1060–1066-
dc.identifier.urihttp://hdl.handle.net/10722/287776-
dc.descriptionShort Paper - Gather Session 1A: Machine Translation and Multilinguality-
dc.description.abstractModern neural machine translation (NMT) models employ a large number of parameters, which leads to serious over-parameterization and typically causes the underutilization of computational resources. In response to this problem, we empirically investigate whether the redundant parameters can be reused to achieve better performance. Experiments and analyses are systematically conducted on different datasets and NMT architectures. We show that: 1) the pruned parameters can be rejuvenated to improve the baseline model by up to +0.8 BLEU points; 2) the rejuvenated parameters are reallocated to enhance the ability of modeling low-level lexical information.-
dc.languageeng-
dc.publisherAssociation for Computational Linguistics.-
dc.relation.ispartofConference on Empirical Methods in Natural Language Processing (EMNLP) 2020-
dc.titleOn the sparsity of neural machine translation models-
dc.typeConference_Paper-
dc.identifier.emailLi, VOK: vli@eee.hku.hk-
dc.identifier.authorityLi, VOK=rp00150-
dc.description.naturelink_to_OA_fulltext-
dc.identifier.doi10.18653/v1/2020.emnlp-main.78-
dc.identifier.hkuros315139-
dc.identifier.spage1060-
dc.identifier.epage1066-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats