TY - JOUR
T1 - Enhancing emotion recognition using multimodal fusion of physiological, environmental, personal data
AU - Kim, Hakpyeong
AU - Hong, Taehoon
N1 - Publisher Copyright:
© 2024 Elsevier Ltd
PY - 2024/9/1
Y1 - 2024/9/1
N2 - Human emotion recognition, crucial for interpersonal relations and human-building interaction, identifies emotions from various behavioral signals to improve user interactions. To enhance the performance of emotion recognition, this study proposed a novel model that fuses physiological, environmental, and personal data. A unique dataset was created via experiments conducted in an environmental chamber, and an emotion recognition model was subsequently developed using a multimodal fusion approach. The model transforms physiological data into 2D images to capture time series and spatial features and uniquely incorporates metadata, including environmental and personal data. The model's generalizability was validated using a leave-one-sample-out approach. The result showed 31.6% reduction of error with a predicted area when physiological, environmental, and personal data were fused in the emotion recognition model, suggesting that incorporating various contextual factors beyond physiological changes, such as the surrounding environment and inherent or acquired individual traits, can significantly enhance the model's understanding of emotions. Furthermore, the model was to be robust to individual differences, offering consistent emotion recognition across different subjects. These findings suggest that the proposed model can serve as a potent tool for emotion recognition in built environmental applications.
AB - Human emotion recognition, crucial for interpersonal relations and human-building interaction, identifies emotions from various behavioral signals to improve user interactions. To enhance the performance of emotion recognition, this study proposed a novel model that fuses physiological, environmental, and personal data. A unique dataset was created via experiments conducted in an environmental chamber, and an emotion recognition model was subsequently developed using a multimodal fusion approach. The model transforms physiological data into 2D images to capture time series and spatial features and uniquely incorporates metadata, including environmental and personal data. The model's generalizability was validated using a leave-one-sample-out approach. The result showed 31.6% reduction of error with a predicted area when physiological, environmental, and personal data were fused in the emotion recognition model, suggesting that incorporating various contextual factors beyond physiological changes, such as the surrounding environment and inherent or acquired individual traits, can significantly enhance the model's understanding of emotions. Furthermore, the model was to be robust to individual differences, offering consistent emotion recognition across different subjects. These findings suggest that the proposed model can serve as a potent tool for emotion recognition in built environmental applications.
KW - Affective computing
KW - Emotion recognition
KW - Human-computer interaction
KW - Multimodal fusion
KW - Swin transformer
UR - http://www.scopus.com/inward/record.url?scp=85188634661&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85188634661&partnerID=8YFLogxK
U2 - 10.1016/j.eswa.2024.123723
DO - 10.1016/j.eswa.2024.123723
M3 - Article
AN - SCOPUS:85188634661
SN - 0957-4174
VL - 249
JO - Expert Systems with Applications
JF - Expert Systems with Applications
M1 - 123723
ER -