Date of Original Version
Spoken Dialogue Systems Technology and Design
The final publication is available at Springer via http://dx.doi.org/10.1007/978-1-4419-7934-6_4
Abstract or Description
Anger recognition in speech dialogue systems can help to enhance human commputer interaction. In this chapter we report on the setup and performance opti-izationtechniques for successful anger classification using acoustic cues. We evaluate the performance of a broad variety of features on both a German and an American English voice portal database which contain “real” (i.e. non-acted) continuous speech of narrow-band quality. Starting with a large-scale feature extraction, we determine optimal sets of feature combinations for each language, by applying an Information-Gain based ranking scheme. Analyzing the ranking we notice that a large proportion of the most promising features for both databases are derived from MFCC and loudness. In contrast to this similarity also pitch features proved importance for the English database. We further calculate classification scores for our setups using discriminative training and Support-Vector Machine classification. The developed systems show that anger
Spoken Dialogue Systems Technology and Design, 83-105.