File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/TETCI.2024.3372440
- Scopus: eid_2-s2.0-105006878736
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Article: Prompt-Based Out-of-Distribution Intent Detection
| Title | Prompt-Based Out-of-Distribution Intent Detection |
|---|---|
| Authors | |
| Keywords | Natural language processing out-of-distribution detection prompt-based learning text classification |
| Issue Date | 1-Jun-2025 |
| Citation | IEEE Transactions on Emerging Topics in Computational Intelligence, 2025, v. 9, n. 3, p. 2371-2382 How to Cite? |
| Abstract | Recent rapid advances in pre-trained language models, such as BERT and GPT, in natural language processing (NLP) have greatly improved the efficacy of text classifiers, easily surpassing human level performance in standard datasets like GLUE. However, most of these standard tasks implicitly assume a closed-world situation, where all testing data are supposed to lie in the same scope or distribution of the training data. Out-of-distribution (OOD) detection is the task of detecting when an input data point lies beyond the scope of the seen training set. This is becoming increasingly important as NLP agents, such as chatbots or virtual assistants, have been being deployed ubiquitously in our daily lives, thus attracting more attention from the research community to make it more accurate and robust at the same time. Recent work can be broadly categorized into two orthogonal approaches - data generative/augmentative methods and threshold/boundary learning. In this work, we follow the former and propose a method for the task based on prompting, which is known for its zero and few-shot capabilities. Generating synthetic outliers in terms of prompts allows the model to more efficiently learn OOD samples than the existing methods. Testing on nine different settings across three standard datasets used for OOD detection, our method with adaptive decision boundary is able to achieve competitive or superior performances compared with the current state-of-the-art in all cases. We also provide extensive analysis on each dataset as well as perform comprehensive ablation studies on each component of our model. |
| Persistent Identifier | http://hdl.handle.net/10722/366996 |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Chow, Rudolf | - |
| dc.contributor.author | Lam, Albert YS | - |
| dc.date.accessioned | 2025-11-29T00:35:48Z | - |
| dc.date.available | 2025-11-29T00:35:48Z | - |
| dc.date.issued | 2025-06-01 | - |
| dc.identifier.citation | IEEE Transactions on Emerging Topics in Computational Intelligence, 2025, v. 9, n. 3, p. 2371-2382 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/366996 | - |
| dc.description.abstract | Recent rapid advances in pre-trained language models, such as BERT and GPT, in natural language processing (NLP) have greatly improved the efficacy of text classifiers, easily surpassing human level performance in standard datasets like GLUE. However, most of these standard tasks implicitly assume a closed-world situation, where all testing data are supposed to lie in the same scope or distribution of the training data. Out-of-distribution (OOD) detection is the task of detecting when an input data point lies beyond the scope of the seen training set. This is becoming increasingly important as NLP agents, such as chatbots or virtual assistants, have been being deployed ubiquitously in our daily lives, thus attracting more attention from the research community to make it more accurate and robust at the same time. Recent work can be broadly categorized into two orthogonal approaches - data generative/augmentative methods and threshold/boundary learning. In this work, we follow the former and propose a method for the task based on prompting, which is known for its zero and few-shot capabilities. Generating synthetic outliers in terms of prompts allows the model to more efficiently learn OOD samples than the existing methods. Testing on nine different settings across three standard datasets used for OOD detection, our method with adaptive decision boundary is able to achieve competitive or superior performances compared with the current state-of-the-art in all cases. We also provide extensive analysis on each dataset as well as perform comprehensive ablation studies on each component of our model. | - |
| dc.language | eng | - |
| dc.relation.ispartof | IEEE Transactions on Emerging Topics in Computational Intelligence | - |
| dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
| dc.subject | Natural language processing | - |
| dc.subject | out-of-distribution detection | - |
| dc.subject | prompt-based learning | - |
| dc.subject | text classification | - |
| dc.title | Prompt-Based Out-of-Distribution Intent Detection | - |
| dc.type | Article | - |
| dc.identifier.doi | 10.1109/TETCI.2024.3372440 | - |
| dc.identifier.scopus | eid_2-s2.0-105006878736 | - |
| dc.identifier.volume | 9 | - |
| dc.identifier.issue | 3 | - |
| dc.identifier.spage | 2371 | - |
| dc.identifier.epage | 2382 | - |
| dc.identifier.eissn | 2471-285X | - |
| dc.identifier.issnl | 2471-285X | - |
