To Örebro University

oru.seÖrebro University Publications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
CN-waterfall: a deep convolutional neural network for multimodal physiological affect detection
Department of Computing Science, Umeå University, Umeå, Sweden.
Örebro University, School of Science and Technology. (Centre for Applied Autonomous Sensor Systems (AASS))ORCID iD: 0000-0002-4001-2087
Department of Computing Science, Umeå University, Umeå, Sweden; School of Science and Technology, Aalto University, Espoo, Finland.
2022 (English)In: Neural Computing & Applications, ISSN 0941-0643, E-ISSN 1433-3058, Vol. 34, no 3, p. 2157-2176Article in journal (Refereed) Published
Abstract [en]

Affective computing solutions, in the literature, mainly rely on machine learning methods designed to accurately detect human affective states. Nevertheless, many of the proposed methods are based on handcrafted features, requiring sufficient expert knowledge in the realm of signal processing. With the advent of deep learning methods, attention has turned toward reduced feature engineering and more end-to-end machine learning. However, most of the proposed models rely on late fusion in a multimodal context. Meanwhile, addressing interrelations between modalities for intermediate-level data representation has been largely neglected. In this paper, we propose a novel deep convolutional neural network, called CN-Waterfall, consisting of two modules: Base and General. While the Base module focuses on the low-level representation of data from each single modality, the General module provides further information, indicating relations between modalities in the intermediate- and high-level data representations. The latter module has been designed based on theoretically grounded concepts in the Explainable AI (XAI) domain, consisting of four different fusions. These fusions are mainly tailored to correlation- and non-correlation-based modalities. To validate our model, we conduct an exhaustive experiment on WESAD and MAHNOB-HCI, two publicly and academically available datasets in the context of multimodal affective computing. We demonstrate that our proposed model significantly improves the performance of physiological-based multimodal affect detection.

Place, publisher, year, edition, pages
Springer, 2022. Vol. 34, no 3, p. 2157-2176
Keywords [en]
Multimodal affect detection, Deep convolutional neural network, Physiological-based sensors, Data fusion
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:oru:diva-94919DOI: 10.1007/s00521-021-06516-3ISI: 000698886400003Scopus ID: 2-s2.0-85115620535OAI: oai:DiVA.org:oru-94919DiVA, id: diva2:1602664
Funder
Knut and Alice Wallenberg Foundation
Note

Funding agency:

Umeå University

Available from: 2021-10-13 Created: 2021-10-13 Last updated: 2022-09-12Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Alirezaie, Marjan

Search in DiVA

By author/editor
Alirezaie, Marjan
By organisation
School of Science and Technology
In the same journal
Neural Computing & Applications
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 100 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf