Show simple item record

dc.contributor.authorMattern, Enni
dc.contributor.authorJackson, Roxanne R.
dc.contributor.authorDoshmanziari, Roya
dc.contributor.authorDewitte, Marieke
dc.contributor.authorVaragnolo, Damiano
dc.contributor.authorKnorn, Steffi
dc.date.accessioned2024-01-12T08:38:16Z
dc.date.available2024-01-12T08:38:16Z
dc.date.created2023-12-18T14:20:42Z
dc.date.issued2023
dc.identifier.citationBioengineering. 2023, 10 (11), .en_US
dc.identifier.issn2306-5354
dc.identifier.urihttps://hdl.handle.net/11250/3111247
dc.description.abstractImplementing affective engineering in real-life applications requires the ability to effectively recognize emotions using physiological measurements. Despite being a widely researched topic, there seems to be a lack of systems that translate results from data collected in a laboratory setting to higher technology readiness levels. In this paper, we delve into the feasibility of emotion recognition beyond controlled laboratory environments. For this reason, we create a minimally-invasive experimental setup by combining emotional recall via autobiographical emotion memory tasks with a user-friendly Empatica wristband measuring blood volume pressure, electrodermal activity, skin temperature, and acceleration. We employ standard practices of feature-based supervised learning and specifically use support vector machines to explore subject dependency through various segmentation methods. We collected data from 45 participants. After preprocessing, using a data set of 134 segments from 40 participants, the accuracy of the classifier after 10-fold cross-validation was barely better than random guessing (36% for four emotions). However, when extracting multiple segments from each emotion task per participant using 10-fold cross-validation (i.e., including subject-dependent data in the training set), the classification rate increased to up to 75% for four emotions but was still as low as 32% for leave-one-subject-out cross-validation (i.e., subject-independent training). We conclude that highly subject-dependent issues might pose emotion recognition.en_US
dc.language.isoengen_US
dc.publisherMDPIen_US
dc.rightsNavngivelse 4.0 Internasjonal*
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/deed.no*
dc.titleEmotion Recognition from Physiological Signals Collected with a Wrist Device and Emotional Recallen_US
dc.title.alternativeEmotion Recognition from Physiological Signals Collected with a Wrist Device and Emotional Recallen_US
dc.typePeer revieweden_US
dc.typeJournal articleen_US
dc.description.versionpublishedVersionen_US
dc.source.pagenumber0en_US
dc.source.volume10en_US
dc.source.journalBioengineeringen_US
dc.source.issue11en_US
dc.identifier.doi10.3390/bioengineering10111308
dc.identifier.cristin2214967
cristin.ispublishedtrue
cristin.fulltextoriginal
cristin.qualitycode1


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

Navngivelse 4.0 Internasjonal
Except where otherwise noted, this item's license is described as Navngivelse 4.0 Internasjonal