This paper describes a system to detect angry vs. non-angry utterances of children who are engaged in dialog with an Aibo robot dog. The system was submitted to the Interspeech2009 Emotion Challenge evaluation. The speech data consist of short utterances of the children�s speech, and the proposed system is designed to detect anger in each given chunk. Frame-based cepstral features, prosodic and acoustic features as well as glottal excitation features are extracted automatically, reduced in dimensionality and classified by means of an artificial neural network and a support vector machine. An automatic speech recognizer transcribes the words in an utterance and yields a separate classification based on the degree of emotional salience of the words. Late fusion is applied to make a final decision on anger vs. non-anger of the utterance. Preliminary results show 75.9% unweighted average recall on the training data and 67.6% on the test set.
|Title of host publication||10th Annual Conference of the International Speech Communication Association (Interspeech 2009)|
|Place of Publication||Brighton, UK|
|Publisher||International Speech Communication Association|
|Number of pages||4|
|Publication status||Published - 2009|
|Event||Interspeech-2009 - Brighton, United Kingdom|
Duration: 6 Sep 2009 → 9 Sep 2009
|Period||6/09/09 → 9/09/09|
Polzehl, T., Sundaram, S., Ketabdar, H., Wagner, M., & Metze, F. (2009). Emotion Classification in Children's speech using fusion of acoustic and linguistic features. In M. Uther (Ed.), 10th Annual Conference of the International Speech Communication Association (Interspeech 2009) (pp. 340-343). Brighton, UK: International Speech Communication Association.