File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Prompt-Based Out-of-Distribution Intent Detection

TitlePrompt-Based Out-of-Distribution Intent Detection
Authors
KeywordsNatural language processing
out-of-distribution detection
prompt-based learning
text classification
Issue Date1-Jun-2025
Citation
IEEE Transactions on Emerging Topics in Computational Intelligence, 2025, v. 9, n. 3, p. 2371-2382 How to Cite?
AbstractRecent rapid advances in pre-trained language models, such as BERT and GPT, in natural language processing (NLP) have greatly improved the efficacy of text classifiers, easily surpassing human level performance in standard datasets like GLUE. However, most of these standard tasks implicitly assume a closed-world situation, where all testing data are supposed to lie in the same scope or distribution of the training data. Out-of-distribution (OOD) detection is the task of detecting when an input data point lies beyond the scope of the seen training set. This is becoming increasingly important as NLP agents, such as chatbots or virtual assistants, have been being deployed ubiquitously in our daily lives, thus attracting more attention from the research community to make it more accurate and robust at the same time. Recent work can be broadly categorized into two orthogonal approaches - data generative/augmentative methods and threshold/boundary learning. In this work, we follow the former and propose a method for the task based on prompting, which is known for its zero and few-shot capabilities. Generating synthetic outliers in terms of prompts allows the model to more efficiently learn OOD samples than the existing methods. Testing on nine different settings across three standard datasets used for OOD detection, our method with adaptive decision boundary is able to achieve competitive or superior performances compared with the current state-of-the-art in all cases. We also provide extensive analysis on each dataset as well as perform comprehensive ablation studies on each component of our model.
Persistent Identifierhttp://hdl.handle.net/10722/366996

 

DC FieldValueLanguage
dc.contributor.authorChow, Rudolf-
dc.contributor.authorLam, Albert YS-
dc.date.accessioned2025-11-29T00:35:48Z-
dc.date.available2025-11-29T00:35:48Z-
dc.date.issued2025-06-01-
dc.identifier.citationIEEE Transactions on Emerging Topics in Computational Intelligence, 2025, v. 9, n. 3, p. 2371-2382-
dc.identifier.urihttp://hdl.handle.net/10722/366996-
dc.description.abstractRecent rapid advances in pre-trained language models, such as BERT and GPT, in natural language processing (NLP) have greatly improved the efficacy of text classifiers, easily surpassing human level performance in standard datasets like GLUE. However, most of these standard tasks implicitly assume a closed-world situation, where all testing data are supposed to lie in the same scope or distribution of the training data. Out-of-distribution (OOD) detection is the task of detecting when an input data point lies beyond the scope of the seen training set. This is becoming increasingly important as NLP agents, such as chatbots or virtual assistants, have been being deployed ubiquitously in our daily lives, thus attracting more attention from the research community to make it more accurate and robust at the same time. Recent work can be broadly categorized into two orthogonal approaches - data generative/augmentative methods and threshold/boundary learning. In this work, we follow the former and propose a method for the task based on prompting, which is known for its zero and few-shot capabilities. Generating synthetic outliers in terms of prompts allows the model to more efficiently learn OOD samples than the existing methods. Testing on nine different settings across three standard datasets used for OOD detection, our method with adaptive decision boundary is able to achieve competitive or superior performances compared with the current state-of-the-art in all cases. We also provide extensive analysis on each dataset as well as perform comprehensive ablation studies on each component of our model.-
dc.languageeng-
dc.relation.ispartofIEEE Transactions on Emerging Topics in Computational Intelligence-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectNatural language processing-
dc.subjectout-of-distribution detection-
dc.subjectprompt-based learning-
dc.subjecttext classification-
dc.titlePrompt-Based Out-of-Distribution Intent Detection-
dc.typeArticle-
dc.identifier.doi10.1109/TETCI.2024.3372440-
dc.identifier.scopuseid_2-s2.0-105006878736-
dc.identifier.volume9-
dc.identifier.issue3-
dc.identifier.spage2371-
dc.identifier.epage2382-
dc.identifier.eissn2471-285X-
dc.identifier.issnl2471-285X-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats