Towards a better gold standard: Denoising and modelling continuous emotion annotations based on feature agglomeration and outlier regularisation

Abstract

Emotions are often perceived by humans through a series of multimodal cues, such as verbal expressions, facial expressions and gestures. In order to recognise emotions automatically, reliable emotional labels are required to learn a mapping from human expressions to corresponding emotions. Dimensional emotion models have become popular and have been widely applied for annotating emotions continuously in the time domain. However, the statistical relationship between emotional dimensions is rarely studied. This paper provides a solution to automatic emotion recognition for the Audio/Visual Emotion Challenge (AVEC) 2018. The objective is to find a robust way to detect emotions using more reliable emotion annotations in the valence and arousal dimensions.

Publication
In Proceedings of the 2018 on Audio/Visual Emotion Challenge and Workshop
Click the Cite button above to enable visitors to import publication metadata into their reference management software.
Avatar
Chen Wang
Phd Candidate of Affective Computing

My research interests include impression/emotion recoginition in human-human interaction and human-robot/agent interaction.