File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Improved zero-shot neural machine translation via ignoring spurious correlations

TitleImproved zero-shot neural machine translation via ignoring spurious correlations
Authors
Issue Date2019
PublisherAssociation for Computational Linguistics.
Citation
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL), Florence, Italy, 28 July - 2 August 2019, p. 1258–1268 How to Cite?
AbstractZero-shot translation, translating between language pairs on which a Neural Machine Translation (NMT) system has never been trained, is an emergent property when training the system in multilingual settings. However, naive training for zero-shot NMT easily fails, and is sensitive to hyper-parameter setting. The performance typically lags far behind the more conventional pivot-based approach which translates twice using a third language as a pivot. In this work, we address the degeneracy problem due to capturing spurious correlations by quantitatively analyzing the mutual information between language IDs of the source and decoded sentences. Inspired by this analysis, we propose to use two simple but effective approaches: (1) decoder pre-training; (2) back-translation. These methods show significant improvement (4 22 BLEU points) over the vanilla zero-shot translation on three challenging multilingual datasets, and achieve similar or better results than the pivot-based approach.
DescriptionPoster Session 2B: Machine Translation - no. 7
Persistent Identifierhttp://hdl.handle.net/10722/277807
ISBN

 

DC FieldValueLanguage
dc.contributor.authorGu, J-
dc.contributor.authorWang, Y-
dc.contributor.authorCho, K-
dc.contributor.authorLi, VOK-
dc.date.accessioned2019-10-04T08:01:44Z-
dc.date.available2019-10-04T08:01:44Z-
dc.date.issued2019-
dc.identifier.citationProceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL), Florence, Italy, 28 July - 2 August 2019, p. 1258–1268-
dc.identifier.isbn978-1-950737-48-2-
dc.identifier.urihttp://hdl.handle.net/10722/277807-
dc.descriptionPoster Session 2B: Machine Translation - no. 7-
dc.description.abstractZero-shot translation, translating between language pairs on which a Neural Machine Translation (NMT) system has never been trained, is an emergent property when training the system in multilingual settings. However, naive training for zero-shot NMT easily fails, and is sensitive to hyper-parameter setting. The performance typically lags far behind the more conventional pivot-based approach which translates twice using a third language as a pivot. In this work, we address the degeneracy problem due to capturing spurious correlations by quantitatively analyzing the mutual information between language IDs of the source and decoded sentences. Inspired by this analysis, we propose to use two simple but effective approaches: (1) decoder pre-training; (2) back-translation. These methods show significant improvement (4 22 BLEU points) over the vanilla zero-shot translation on three challenging multilingual datasets, and achieve similar or better results than the pivot-based approach.-
dc.languageeng-
dc.publisherAssociation for Computational Linguistics.-
dc.relation.ispartofProceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL)-
dc.titleImproved zero-shot neural machine translation via ignoring spurious correlations-
dc.typeConference_Paper-
dc.identifier.emailWang, Y: wangyong@eee.hku.hk-
dc.identifier.emailLi, VOK: vli@eee.hku.hk-
dc.identifier.authorityLi, VOK=rp00150-
dc.identifier.doi10.18653/v1/P19-1121-
dc.identifier.hkuros306515-
dc.identifier.spage1258-
dc.identifier.epage1268-
dc.publisher.placeStroudsburg, PA-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats