File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.3390/electronics14214304
- Find via

Supplementary
-
Citations:
- Appears in Collections:
Article: An Emotional AI Chatbot Using an Ontology and a Novel Audiovisual Emotion Transformer for Improving Nonverbal Communication
| Title | An Emotional AI Chatbot Using an Ontology and a Novel Audiovisual Emotion Transformer for Improving Nonverbal Communication |
|---|---|
| Authors | |
| Issue Date | 31-Oct-2025 |
| Publisher | MDPI |
| Citation | Electronics, 2025, v. 14, n. 21 How to Cite? |
| Abstract | One of the key limitations of AI chatbots is the lack of human-like nonverbal communication. Although there are many research studies on video or audio emotion recognition for detecting human emotions, there is no research that combines video, audio, and ontology methods to develop an AI chatbot with human-like communication. Therefore, this research aims to develop an audio-video emotion recognition model and an emotion-ontology-based chatbot engine to improve human-like communication with emotion detection. This research proposed a novel model of cluster-based audiovisual emotion recognition for improving emotion detection with both video and audio signals and compared it with existing methods using video or audio signals only. Twenty-two audio features, the Mel spectrogram, and facial action units were extracted, and the last two were fed into a cluster-based independent transformer to learn long-term temporal dependencies. Our model was validated on three public audiovisual datasets: RAVDESS, SAVEE, and RML. The results demonstrated that the accuracy scores of the clustered transformer model for RAVDESS, SAVEE, and RML were 86.46%, 92.71%, and 91.67%, respectively, outperforming the existing best model with accuracy scores of 86.3%, 75%, and 60.2%, respectively. An emotion-ontology-based chatbot engine was implemented to make inquiry responses based on the detected emotion. A case study of the HKU Campusland metaverse was used as proof of concept of the emotional AI chatbot for nonverbal communication. |
| Persistent Identifier | http://hdl.handle.net/10722/365958 |
| ISSN | 2023 Impact Factor: 2.6 2023 SCImago Journal Rankings: 0.644 |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Wang, Yun | - |
| dc.contributor.author | Cheung, Liege | - |
| dc.contributor.author | Ma, Patrick | - |
| dc.contributor.author | Lee, Herbert | - |
| dc.contributor.author | Lau, Adela S.M. | - |
| dc.date.accessioned | 2025-11-14T02:40:40Z | - |
| dc.date.available | 2025-11-14T02:40:40Z | - |
| dc.date.issued | 2025-10-31 | - |
| dc.identifier.citation | Electronics, 2025, v. 14, n. 21 | - |
| dc.identifier.issn | 2079-9292 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/365958 | - |
| dc.description.abstract | <p>One of the key limitations of AI chatbots is the lack of human-like nonverbal communication. Although there are many research studies on video or audio emotion recognition for detecting human emotions, there is no research that combines video, audio, and ontology methods to develop an AI chatbot with human-like communication. Therefore, this research aims to develop an audio-video emotion recognition model and an emotion-ontology-based chatbot engine to improve human-like communication with emotion detection. This research proposed a novel model of cluster-based audiovisual emotion recognition for improving emotion detection with both video and audio signals and compared it with existing methods using video or audio signals only. Twenty-two audio features, the Mel spectrogram, and facial action units were extracted, and the last two were fed into a cluster-based independent transformer to learn long-term temporal dependencies. Our model was validated on three public audiovisual datasets: RAVDESS, SAVEE, and RML. The results demonstrated that the accuracy scores of the clustered transformer model for RAVDESS, SAVEE, and RML were 86.46%, 92.71%, and 91.67%, respectively, outperforming the existing best model with accuracy scores of 86.3%, 75%, and 60.2%, respectively. An emotion-ontology-based chatbot engine was implemented to make inquiry responses based on the detected emotion. A case study of the HKU Campusland metaverse was used as proof of concept of the emotional AI chatbot for nonverbal communication.</p> | - |
| dc.language | eng | - |
| dc.publisher | MDPI | - |
| dc.relation.ispartof | Electronics | - |
| dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
| dc.title | An Emotional AI Chatbot Using an Ontology and a Novel Audiovisual Emotion Transformer for Improving Nonverbal Communication | - |
| dc.type | Article | - |
| dc.identifier.doi | 10.3390/electronics14214304 | - |
| dc.identifier.volume | 14 | - |
| dc.identifier.issue | 21 | - |
| dc.identifier.eissn | 2079-9292 | - |
| dc.identifier.issnl | 2079-9292 | - |
