File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
  • Find via Find It@HKUL

Conference Paper: Show Me How To Revise: Improving Lexically Constrained Sentence Generation with XLNet

TitleShow Me How To Revise: Improving Lexically Constrained Sentence Generation with XLNet
Authors
KeywordsGeneration
Applications
Language Models
Issue Date2021
PublisherAAAI Press. The Journal's web site is located at https://aaai.org/Library/AAAI/aaai-library.php
Citation
Proceedings of the 35th Association for the Advancement of Artificial Intelligence (AAAI) Conference on Artificial Intelligence (AAAI-21), Virtual Conference, USA, 2-9 February 2021, v. 35 n. 14, p. 12989-12997 How to Cite?
AbstractLexically constrained sentence generation allows the incorporation of prior knowledge such as lexical constraints into the output. This technique has been applied to machine translation, and dialog response generation. Previous work usually used Markov Chain Monte Carlo (MCMC) sampling to generate lexically constrained sentences, but they randomly determined the position to be edited and the action to be taken, resulting in many invalid refinements. To overcome this challenge, we used a classifier to instruct the MCMC-based models where and how to refine the candidate sentences. First, we developed two methods to create synthetic data on which the pre-trained model is fine-tuned to obtain a reliable classifier. Next, we proposed a two-step approach, “Predict and Revise”, for constrained sentence generation. During the predict step, we leveraged the classifier to compute the learned prior for the candidate sentence. During the revise step, we resorted to MCMC sampling to revise the candidate sentence by conducting a sampled action at a sampled position drawn from the learned prior. We compared our proposed models with many strong baselines on two tasks, generating sentences with lexical constraints and text infilling. Experimental results have demonstrated that our proposed model performs much better than the previous work in terms of sentence fluency and diversity. Our code, pre-trained models and Appendix are available at https://github.com/NLPCode/MCMCXLNet.
DescriptionAAAI-21 Technical Tracks 14 / AAAI Technical Track on Speech and Natural Language Processing I
Persistent Identifierhttp://hdl.handle.net/10722/305497
ISSN

 

DC FieldValueLanguage
dc.contributor.authorHe, X-
dc.contributor.authorLi, VOK-
dc.date.accessioned2021-10-20T10:10:14Z-
dc.date.available2021-10-20T10:10:14Z-
dc.date.issued2021-
dc.identifier.citationProceedings of the 35th Association for the Advancement of Artificial Intelligence (AAAI) Conference on Artificial Intelligence (AAAI-21), Virtual Conference, USA, 2-9 February 2021, v. 35 n. 14, p. 12989-12997-
dc.identifier.issn2159-5399-
dc.identifier.urihttp://hdl.handle.net/10722/305497-
dc.descriptionAAAI-21 Technical Tracks 14 / AAAI Technical Track on Speech and Natural Language Processing I-
dc.description.abstractLexically constrained sentence generation allows the incorporation of prior knowledge such as lexical constraints into the output. This technique has been applied to machine translation, and dialog response generation. Previous work usually used Markov Chain Monte Carlo (MCMC) sampling to generate lexically constrained sentences, but they randomly determined the position to be edited and the action to be taken, resulting in many invalid refinements. To overcome this challenge, we used a classifier to instruct the MCMC-based models where and how to refine the candidate sentences. First, we developed two methods to create synthetic data on which the pre-trained model is fine-tuned to obtain a reliable classifier. Next, we proposed a two-step approach, “Predict and Revise”, for constrained sentence generation. During the predict step, we leveraged the classifier to compute the learned prior for the candidate sentence. During the revise step, we resorted to MCMC sampling to revise the candidate sentence by conducting a sampled action at a sampled position drawn from the learned prior. We compared our proposed models with many strong baselines on two tasks, generating sentences with lexical constraints and text infilling. Experimental results have demonstrated that our proposed model performs much better than the previous work in terms of sentence fluency and diversity. Our code, pre-trained models and Appendix are available at https://github.com/NLPCode/MCMCXLNet.-
dc.languageeng-
dc.publisherAAAI Press. The Journal's web site is located at https://aaai.org/Library/AAAI/aaai-library.php-
dc.relation.ispartofProceedings of the AAAI Conference on Artificial Intelligence-
dc.subjectGeneration-
dc.subjectApplications-
dc.subjectLanguage Models-
dc.titleShow Me How To Revise: Improving Lexically Constrained Sentence Generation with XLNet-
dc.typeConference_Paper-
dc.identifier.emailLi, VOK: vli@eee.hku.hk-
dc.identifier.authorityLi, VOK=rp00150-
dc.identifier.hkuros327679-
dc.identifier.volume35-
dc.identifier.issue14-
dc.identifier.spage12989-
dc.identifier.epage12997-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats