File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1145/3357384.3358160
- Scopus: eid_2-s2.0-85075478890
- WOS: WOS:000539898202014
Supplementary
- Citations:
- Appears in Collections:
Conference Paper: An explainable deep fusion network for affect recognition using physiological signals
| Title | An explainable deep fusion network for affect recognition using physiological signals |
|---|---|
| Authors | |
| Keywords | Affect recognition Deep learning Explainability Multimodal fusion |
| Issue Date | 2019 |
| Citation | International Conference on Information and Knowledge Management, Proceedings, 2019, p. 2069-2072 How to Cite? |
| Abstract | Affective computing is an emerging research area which provides insights on human's mental state through human-machine interaction. During the interaction process, bio-signal analysis is essential to detect human affective changes. Currently, machine learning methods to analyse bio-signals are the state of the art to detect the affective states, but most empirical works mainly deploy traditional machine learning methods rather than deep learning models due to the need for explainability. In this paper, we propose a deep learning model to process multimodal-multisensory bio-signals for affect recognition. It supports batch training for different sampling rate signals at the same time, and our results show significant improvement compared to the state of the art. Furthermore, the results are interpreted at the sensor- and signal- level to improve the explainaibility of our deep learning model. |
| Persistent Identifier | http://hdl.handle.net/10722/354145 |
| ISI Accession Number ID |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Lin, Jionghao | - |
| dc.contributor.author | Pan, Shirui | - |
| dc.contributor.author | Lee, Cheng Siong | - |
| dc.contributor.author | Oviatt, Sharon | - |
| dc.date.accessioned | 2025-02-07T08:46:44Z | - |
| dc.date.available | 2025-02-07T08:46:44Z | - |
| dc.date.issued | 2019 | - |
| dc.identifier.citation | International Conference on Information and Knowledge Management, Proceedings, 2019, p. 2069-2072 | - |
| dc.identifier.uri | http://hdl.handle.net/10722/354145 | - |
| dc.description.abstract | Affective computing is an emerging research area which provides insights on human's mental state through human-machine interaction. During the interaction process, bio-signal analysis is essential to detect human affective changes. Currently, machine learning methods to analyse bio-signals are the state of the art to detect the affective states, but most empirical works mainly deploy traditional machine learning methods rather than deep learning models due to the need for explainability. In this paper, we propose a deep learning model to process multimodal-multisensory bio-signals for affect recognition. It supports batch training for different sampling rate signals at the same time, and our results show significant improvement compared to the state of the art. Furthermore, the results are interpreted at the sensor- and signal- level to improve the explainaibility of our deep learning model. | - |
| dc.language | eng | - |
| dc.relation.ispartof | International Conference on Information and Knowledge Management, Proceedings | - |
| dc.subject | Affect recognition | - |
| dc.subject | Deep learning | - |
| dc.subject | Explainability | - |
| dc.subject | Multimodal fusion | - |
| dc.title | An explainable deep fusion network for affect recognition using physiological signals | - |
| dc.type | Conference_Paper | - |
| dc.description.nature | link_to_subscribed_fulltext | - |
| dc.identifier.doi | 10.1145/3357384.3358160 | - |
| dc.identifier.scopus | eid_2-s2.0-85075478890 | - |
| dc.identifier.spage | 2069 | - |
| dc.identifier.epage | 2072 | - |
| dc.identifier.isi | WOS:000539898202014 | - |
