An explainable deep fusion network for affect recognition using physiological signals
- Publisher:
- ACM
- Publication Type:
- Conference Proceeding
- Citation:
- International Conference on Information and Knowledge Management, Proceedings, 2019, pp. 2069-2072
- Issue Date:
- 2019-11-03
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
3357384.3358160.pdf | Published version | 1.1 MB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
© 2019 Copyright held by the owner/author(s). Publication rights licensed to ACM. Affective computing is an emerging research area which provides insights on human's mental state through human-machine interaction. During the interaction process, bio-signal analysis is essential to detect human affective changes. Currently, machine learning methods to analyse bio-signals are the state of the art to detect the affective states, but most empirical works mainly deploy traditional machine learning methods rather than deep learning models due to the need for explainability. In this paper, we propose a deep learning model to process multimodal-multisensory bio-signals for affect recognition. It supports batch training for different sampling rate signals at the same time, and our results show significant improvement compared to the state of the art. Furthermore, the results are interpreted at the sensor- and signal- level to improve the explainaibility of our deep learning model.
Please use this identifier to cite or link to this item: