File Download
Supplementary

postgraduate thesis: Lexically constrained text generation

TitleLexically constrained text generation
Authors
Advisors
Advisor(s):Yiu, SM
Issue Date2023
PublisherThe University of Hong Kong (Pokfulam, Hong Kong)
Citation
He, X. [贺星伟]. (2023). Lexically constrained text generation. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.
AbstractTeaching machines to generate high-quality natural texts comparable to human writing, has been a longstanding challenge in natural language processing. For a long time, statistical methods dominated natural language generation. In the past decade, we have witnessed many landmark developments in neural text generation, such as the encoder-decoder structure, attention mechanism, transformer and large-scale pretrained language models, which have made neural text generation far superior to statistical text generation. Pre-trained language models, such as BART and GPT-3, have garnered attention from numerous researchers in recent years due to their remarkable performance in various natural language generation tasks. Nowadays, pre-trained language models have become the standard paradigm for text generation. Unfortunately, even though large-scale pre-trained language models have shown promising text generation capabilities, controlling the attributes of the generated text, which is usually referred to as controllable text generation, is not an easy task for users. The controllable aspects of text generation can take different forms, ranging from text sentiments (such as positive, negative, and neutral), text topics (such as sports, entertainment, and politics), text formats (such as poems and couplets), or even the identity of the person writing the text (such as gender and age). In this thesis, we focus on lexically constrained text generation, which falls under controllable text generation. The task involves incorporating pre-specified keywords into generated outputs while maintaining the generation quality. The ability to generate text that meets these criteria has broad implications for downstream tasks, including but not limited to generating dialog responses, crafting stories, composing product advertisements, and creating meeting summaries based on key phrases. This thesis proposes models to improve the generation quality and reduce the inference latency for lexically constrained text generation. We compare our approaches with previous models on different datasets, such as One-Billion-Word, Yelp, CommonGen and Oxford. Extensive experiment results demonstrate the effectiveness of our proposed methods. The contributions of this thesis to lexically constrained text generation include: (1) designing a two-step approach, “Predict and Revise”, which improves the generation quality by introducing a predictor to guide the model to refine the candidate outputs; (2) inventing Constrained BART (CBART), which accelerates the inference process by refining multiple tokens of the candidate output in parallel and improves the generation quality with the pre-trained model, BART; (3) proposing metric-guided distillation, which enables the retriever and ranker to select more relevant sentences by distilling knowledge from the metric to the ranker and retriever, and applying it to the retrieve-and-generate pipeline for commonsense generation (CommonGen), a much more challenging task related to lexically constrained text generation; (4) introducing dictionary example sentence generation, a useful application related to lexically constrained text generation, developing a controllable target-word-aware model and several baselines, releasing a new dataset, proposing two automatic evaluation metrics for this task, and exploring how to control the readability of the generated examples.
DegreeDoctor of Philosophy
SubjectNatural language generation (Computer science)
Dept/ProgramComputer Science
Persistent Identifierhttp://hdl.handle.net/10722/328567

 

DC FieldValueLanguage
dc.contributor.advisorYiu, SM-
dc.contributor.authorHe, Xingwei-
dc.contributor.author贺星伟-
dc.date.accessioned2023-06-29T05:44:17Z-
dc.date.available2023-06-29T05:44:17Z-
dc.date.issued2023-
dc.identifier.citationHe, X. [贺星伟]. (2023). Lexically constrained text generation. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.-
dc.identifier.urihttp://hdl.handle.net/10722/328567-
dc.description.abstractTeaching machines to generate high-quality natural texts comparable to human writing, has been a longstanding challenge in natural language processing. For a long time, statistical methods dominated natural language generation. In the past decade, we have witnessed many landmark developments in neural text generation, such as the encoder-decoder structure, attention mechanism, transformer and large-scale pretrained language models, which have made neural text generation far superior to statistical text generation. Pre-trained language models, such as BART and GPT-3, have garnered attention from numerous researchers in recent years due to their remarkable performance in various natural language generation tasks. Nowadays, pre-trained language models have become the standard paradigm for text generation. Unfortunately, even though large-scale pre-trained language models have shown promising text generation capabilities, controlling the attributes of the generated text, which is usually referred to as controllable text generation, is not an easy task for users. The controllable aspects of text generation can take different forms, ranging from text sentiments (such as positive, negative, and neutral), text topics (such as sports, entertainment, and politics), text formats (such as poems and couplets), or even the identity of the person writing the text (such as gender and age). In this thesis, we focus on lexically constrained text generation, which falls under controllable text generation. The task involves incorporating pre-specified keywords into generated outputs while maintaining the generation quality. The ability to generate text that meets these criteria has broad implications for downstream tasks, including but not limited to generating dialog responses, crafting stories, composing product advertisements, and creating meeting summaries based on key phrases. This thesis proposes models to improve the generation quality and reduce the inference latency for lexically constrained text generation. We compare our approaches with previous models on different datasets, such as One-Billion-Word, Yelp, CommonGen and Oxford. Extensive experiment results demonstrate the effectiveness of our proposed methods. The contributions of this thesis to lexically constrained text generation include: (1) designing a two-step approach, “Predict and Revise”, which improves the generation quality by introducing a predictor to guide the model to refine the candidate outputs; (2) inventing Constrained BART (CBART), which accelerates the inference process by refining multiple tokens of the candidate output in parallel and improves the generation quality with the pre-trained model, BART; (3) proposing metric-guided distillation, which enables the retriever and ranker to select more relevant sentences by distilling knowledge from the metric to the ranker and retriever, and applying it to the retrieve-and-generate pipeline for commonsense generation (CommonGen), a much more challenging task related to lexically constrained text generation; (4) introducing dictionary example sentence generation, a useful application related to lexically constrained text generation, developing a controllable target-word-aware model and several baselines, releasing a new dataset, proposing two automatic evaluation metrics for this task, and exploring how to control the readability of the generated examples.-
dc.languageeng-
dc.publisherThe University of Hong Kong (Pokfulam, Hong Kong)-
dc.relation.ispartofHKU Theses Online (HKUTO)-
dc.rightsThe author retains all proprietary rights, (such as patent rights) and the right to use in future works.-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subject.lcshNatural language generation (Computer science)-
dc.titleLexically constrained text generation-
dc.typePG_Thesis-
dc.description.thesisnameDoctor of Philosophy-
dc.description.thesislevelDoctoral-
dc.description.thesisdisciplineComputer Science-
dc.description.naturepublished_or_final_version-
dc.date.hkucongregation2023-
dc.identifier.mmsid991044695780503414-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats