RESEARCH

HOME RESEARCH
Behavior Computing
States and Traits
Other: Signal Modeling for Understanding
A thin-slice perception of emotion? An information theoretic-based framework to identify locally emotion-rich behavior segments for global affect recognition
Abstract
Human's judgment has been shown to be thin-sliced in nature, i.e., accurate perception can often be achieved for a short duration of exposure to expressive behaviors. In this work, we develop a mutual information-based framework to select the most emotion-rich 20% of local multimodal behavior segments within a 3-minute long affective dyadic interaction in the USC CreativeIT database. We obtain a prediction accuracy of 0.597, 0.728, and 0.772 (measured by Spearman correlation) for an actor's global (session-level) emotion attributes (activation, dominance, and valence) using Fishervector encoding and support vector regression built on these 20% of multimodal emotion-rich behavior segments. Our framework achieves a better accuracy over using the interaction in its entirety and a variety of other data selection baseline methods by a significant margin. Furthermore, our analysis indicates that the highest prediction accuracy can be obtained using only 20% - 30% of data within each session, i.e., additional evidences for the thin-slice nature of affect perception.
Figures
A complete workflow of our mutual information-based framework for identifying locally emotion-rich behavior segments as our thin slice representation
A complete workflow of our mutual information-based framework for identifying locally emotion-rich behavior segments as our thin slice representation
Keywords
thin-slice theory | behavioral signal processing | emotion recognition | multimodal signal processing
Authors
Chi-Chun Lee
Publication Date
2016/03/20
Conference
IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
DOI
10.1109/icassp.2016.7472787
Publisher
IEEE