Date of Original Version

11-2010

Type

Book Chapter

Journal Title

Spoken Dialogue Systems Technology and Design

First Page

83

Last Page

105

Rights Management

The final publication is available at Springer via http://dx.doi.org/10.1007/978-1-4419-7934-6_4

Abstract or Description

Anger recognition in speech dialogue systems can help to enhance human commputer interaction. In this chapter we report on the setup and performance opti-izationtechniques for successful anger classification using acoustic cues. We evaluate the performance of a broad variety of features on both a German and an American English voice portal database which contain “real” (i.e. non-acted) continuous speech of narrow-band quality. Starting with a large-scale feature extraction, we determine optimal sets of feature combinations for each language, by applying an Information-Gain based ranking scheme. Analyzing the ranking we notice that a large proportion of the most promising features for both databases are derived from MFCC and loudness. In contrast to this similarity also pitch features proved importance for the English database. We further calculate classification scores for our setups using discriminative training and Support-Vector Machine classification. The developed systems show that anger

DOI

10.1007/978-1-4419-7934-6_4

Share

COinS
 

Published In

Spoken Dialogue Systems Technology and Design, 83-105.