Sensors | |
CorrNet: Fine-Grained Emotion Recognition for Video Watching Using Wearable Physiological Sensors | |
AbdallahEl Ali1  Chen Wang2  Tianyi Zhang3  Alan Hanjalic3  Pablo Cesar3  | |
[1] Centrum Wiskunde & Informatica (CWI), 1098XG Amsterdam, The Netherlands;Future Media and Convergence Institute, Xinhuanet & State Key Laboratory of Media Convergence Production Technology and Systems, Xinhua News Agency, Beijing 100000, China;Multimedia Computing Group, Delft University of Technology, 2600AA Delft, The Netherlands; | |
关键词: emotion recognition; video; physiological signals; machine learning; | |
DOI : 10.3390/s21010052 | |
来源: DOAJ |
【 摘 要 】
Recognizing user emotions while they watch short-form videos anytime and anywhere is essential for facilitating video content customization and personalization. However, most works either classify a single emotion per video stimuli, or are restricted to static, desktop environments. To address this, we propose a correlation-based emotion recognition algorithm (CorrNet) to recognize the valence and arousal (V-A) of each instance (fine-grained segment of signals) using only wearable, physiological signals (e.g., electrodermal activity, heart rate). CorrNet takes advantage of features both inside each instance (intra-modality features) and between different instances for the same video stimuli (correlation-based features). We first test our approach on an indoor-desktop affect dataset (CASE), and thereafter on an outdoor-mobile affect dataset (MERCA) which we collected using a smart wristband and wearable eyetracker. Results show that for subject-independent binary classification (high-low), CorrNet yields promising recognition accuracies:
【 授权许可】
Unknown