Abstract / Description of output
Recognizing emotional reactions of movie audiences to affective movie content is a challenging task in affective computing. Previous research on induced emotion recognition has mainly focused on using audiovisual movie content. Nevertheless, the relationship between the perceptions of the affective movie content (perceived emotions) and the emotions evoked in the audiences (induced emotions) is unexplored. In this work, we studied the relationship between perceived and induced emotions of movie audiences. Moreover, we investigated multimodal modelling approaches to predict movie induced emotions from movie content based features, as well as physiological and behavioral reactions of movie audiences. To carry out analysis of induced and perceived emotions, we first extended an existing database for movie affect analysis by annotating perceived emotions in a crowd-sourced manner. We find that perceived and induced emotions are not always consistent with each other. In addition, we show that perceived emotions, movie dialogues and aesthetic highlights are discriminative for movie induced emotion recognition besides spectators? physiological and behavioral reactions. Also, our experiments revealed that induced emotion recognition could benefit from including temporal information and performing multimodal fusion. Moreover, our work deeply investigated the gap between affective content analysis and induced emotion recognition by gaining insight into the relationships between aesthetic highlights, induced emotions and perceived emotions.
Original language | English |
---|---|
Number of pages | 17 |
Journal | IEEE Transactions on Affective Computing |
Early online date | 27 Feb 2019 |
DOIs | |
Publication status | E-pub ahead of print - 27 Feb 2019 |
Keywords / Materials (for Non-textual outputs)
- Motion pictures
- Emotion recognition
- Feature extraction
- Physiology
- Databases
- Brain modeling
- Analytical models
- Affective Computing
- Implicit Tagging
- Emotion Recognition
- Multimodal Learning
- Multimodal Fusion
- Induced and Perceived Emotions
- Aesthetic Highlights
- Physiological and Behavioral Signals
- Crowdsourcing
Fingerprint
Dive into the research topics of 'Recognizing Induced Emotions of Movie Audiences From Multimodal Information'. Together they form a unique fingerprint.Profiles
-
Catherine Lai
- School of Philosophy, Psychology and Language Sciences - Lecturer in Speech and Language Processing
- Institute of Language, Cognition and Computation
- Centre for Speech Technology Research
Person: Academic: Research Active