AUTOMATIC SPEECH CLASSIFICATION TO FIVE EMOTIONAL STATES BASED ON GENDER INFORMATION (TuePmPO2)
Author(s) :
Dimitrios Ververidis (Artificial Inteligence and Information Analysis Laboratory, Department of Informatics, Aristotle University of Thessaloniki, Greece)
Constantine Kotropoulos (Artificial Inteligence and Information Analysis Laboratory, Department of Informatics, Aristotle University of Thessaloniki, Greece)
Abstract : Emotional speech recognition aims to automatically classify speech units (e.g., utterances) into emotional states, such as anger, happiness, neutral, sadness and surprise. The major contribution of this paper is to rate the discriminating capability of a set of features for emotional speech recognition when gender information is taken into consideration. A total of 87 features has been calculated over 500 utterances of the Danish Emotional Speech database. The Sequential Forward Selection method (SFS) has been used in order to discover the 5-10 features which are able to classify the samples in the best way for each gender. The criterion used in SFS is the crossvalidated correct classification rate of a Bayes classifier where the class probability distribution functions (pdfs) are approximated via Parzen windows or modeled as Gaussians. When a Bayes classifier with Gaussian pdfs is employed, a correct classification rate of 61.1\% is obtained for male subjects and a corresponding rate of 57.1\% for female ones. In the same experiment, a random classification would result in a correct classification rate of 20\%. When gender information is not considered a correct classification score of 50.6\% is obtained.

Menu