Publication:
Multi-Modal Fusion Learning Through Biosignal, Audio, and Visual Content for Detection of Mental Stress

dc.contributor.authorGÜLİN, DOĞAN
dc.contributor.authorAKBULUT, FATMA PATLAR
dc.date.accessioned2023-10-26T10:58:09Z
dc.date.available2023-10-26T10:58:09Z
dc.date.issued2023
dc.description.abstractMental stress is a significant risk factor for several maladies and can negatively impact a person's quality of life, including their work and personal relationships. Traditional methods of detecting mental stress through interviews and questionnaires may not capture individuals' instantaneous emotional responses. In this study, the method of experience sampling was used to analyze the participants' immediate affective responses, which provides a more comprehensive and dynamic understanding of the participants' experiences. WorkStress3D dataset was compiled using information gathered from 20 participants for three distinct modalities. During an average of one week, 175 h of data containing physiological signals such as BVP, EDA, and body temperature, as well as facial expressions and auditory data, were collected from a single subject. We present a novel fusion model that uses double-early fusion approaches to combine data from multiple modalities. The model's F1 score of 0.94 with a loss of 0.18 is very encouraging, showing that it can accurately identify and classify varying degrees of stress. Furthermore, we investigate the utilization of transfer learning techniques to improve the efficacy of our stress detection system. Despite our efforts, we were unable to attain better results than the fusion model. Transfer learning resulted in an accuracy of 0.93 and a loss of 0.17, illustrating the difficulty of adapting pre-trained models to the task of stress analysis. The results we obtained emphasize the significance of multi-modal fusion in stress detection and the importance of selecting the most suitable model architecture for the given task. The proposed fusion model demonstrates its potential for achieving an accurate and robust classification of stress. This research contributes to the field of stress analysis and contributes to the development of effective models for stress detection.en
dc.description.sponsorshipThis work was supported by the Scientific Research Projects Coordination Unit of Istanbul Kultur University with project number: IKU-BAP2012. Within the scope of the study, data were collected with the permission of the ethics committee of Istanbul Kultur.
dc.identifier.citationDogan, G., Akbulut, F.P. Multi-modal fusion learning through biosignal, audio, and visual content for detection of mental stress. Neural Comput & Applic (2023).
dc.identifier.issn0941-0643
dc.identifier.scopus2-s2.0-85173110378
dc.identifier.urihttps://doi.org/10.1007/s00521-023-09036-4
dc.identifier.urihttps://hdl.handle.net/11413/8847
dc.identifier.wos1077141300005
dc.language.isoen
dc.publisherSpringer London Ltd.
dc.relation.journalNeural Computing & Applications
dc.rightsinfo:eu-repo/semantics/restrictedAccess
dc.subjectStress Detection
dc.subjectSequential and Non-sequential Model
dc.subjectFine-tuning
dc.subjectMulti-modality
dc.titleMulti-Modal Fusion Learning Through Biosignal, Audio, and Visual Content for Detection of Mental Stressen
dc.typeArticle Early Access
dspace.entity.typePublication
local.indexed.atWOS
local.indexed.atScopus
local.journal.endpage20
local.journal.startpage1
relation.isAuthorOfPublication16c815c6-a2cb-439b-b155-9ca020f8cc04
relation.isAuthorOfPublication.latestForDiscovery16c815c6-a2cb-439b-b155-9ca020f8cc04

Files

Original bundle

Now showing 1 - 1 of 1
Placeholder
Name:
↓ Tam Metin/Full Text
Size:
2.05 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Placeholder
Name:
license.txt
Size:
1.81 KB
Format:
Item-specific license agreed upon to submission
Description: