Recognizing child's emotional state in problem-solving child-machine interactions


Yildirim S., Narayanan S.

2nd Workshop on Child, Computer and Interaction, WOCCI '09, Cambridge, MA, Amerika Birleşik Devletleri, 05 Kasım 2009, (Tam Metin Bildiri) identifier

  • Yayın Türü: Bildiri / Tam Metin Bildiri
  • Doi Numarası: 10.1145/1640377.1640391
  • Basıldığı Şehir: Cambridge, MA
  • Basıldığı Ülke: Amerika Birleşik Devletleri
  • Anahtar Kelimeler: Child-computer interaction, Emotion recognition, Spoken dialog systems
  • Hatay Mustafa Kemal Üniversitesi Adresli: Evet

Özet

The need for automatic recognition of a speaker's emotion within a spoken dialog system framework has received increased attention with demand for computer interfaces that provide natural and user-adaptive spoken interaction. This paper addresses the problem of automatically recognizing a child's emotional state using information obtained from audio and video signals. The study is based on a multimodal data corpus consisting of spontaneous conversations between a child and a computer agent. Four different techniques - k-nearest neighborhood (k-NN) classifier, decision tree, linear discriminant classifier (LDC), and support vector machine classifier (SVC) - were employed for classifying utterances into 2 emotion classes, negative and non-negative, for both acoustic and visual information. Experimental results show that, overall, combining visual information with acoustic information leads to performance improvements in emotion recognition. We obtained the best results when information sources were combined at feature level. Specifically, results showed that the addition of visual information to acoustic information yields relative improvements in emotion recognition of 3.8% with both LDC and SVC classifiers for information fusion at the feature level over that of using only acoustic information. Copyright 2009 ACM.