The automatic recognition of emotions in speech

Anton Batliner*, Björn Schuller, Dino Seppi, Stefan Steidl, Laurence Devillers, Laurence Vidrascu, Thurid Vogt, Vered Aharonson, Noam Amir

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingChapterpeer-review


In this chapter, we focus on the automatic recognition of emotional states using acoustic and linguistic parameters as features and classifiers as tools to predict the ‘correct’ emotional states. We first sketch history and state of the art in this field; then we describe the process of ‘corpus engineering’, i.e. the design and the recording of databases, the annotation of emotional states, and further processing such as manual or automatic segmentation. Next, we present an overview of acoustic and linguistic features that are extracted automatically or manually. In the section on classifiers, we deal with topics such as the curse of dimensionality and the sparse data problem, classifiers, and evaluation. At the end of each section, we point out important aspects that should be taken into account for the planning or the assessment of studies. The subject area of this chapter is not emotions in some narrow sense but in a wider sense encompassing emotion-related states such as moods, attitudes, or interpersonal stances as well. We do not aim at an in-depth treatise of some specific aspects or algorithms but at an overview of approaches and strategies that have been used or should be used.

Original languageEnglish
Title of host publicationCognitive Technologies
PublisherSpringer Verlag
Number of pages29
StatePublished - 2011

Publication series

NameCognitive Technologies
ISSN (Print)1611-2482


Dive into the research topics of 'The automatic recognition of emotions in speech'. Together they form a unique fingerprint.

Cite this