Publication Type : Journal Article
Publisher : International Journal of Speech Technology
Source : International Journal of Speech Technology, Volume 22, Number 2, p.327–340 (2019)
Url : https://doi.org/10.1007/s10772-019-09605-w
Campus : Amritapuri
School : School of Engineering
Department : Electronics and Communication
Year : 2019
Abstract : During the past decades, emotion recognition from speech has become one of the most explored areas in affective computing. These systems lack universality due to multilingualism. Research in this direction is restrained due to unavailability of emotional speech databases in various spoken languages. Arabic is one such language, which faces this inadequacy. The proposed work aims at developing a speech emotion recognition system for Arabic speaking community. A speech database with elicited emotions–-anger, happiness, sadness, disgust, surprise and neutrality are recorded from 14 subjects, who are non-native, but proficient speakers in the language. The prosodic, spectral and cepstral features are extracted after pre-processing. Subsequently the features were subjected to single stage classification using supervised learning methods viz. Support vector machine and Extreme learning machine. The performance of the speech emotion recognition systems implemented are compared in terms of accuracy, specificity, precision and recall. Further analysis is carried out by adopting three multistage classification schemes. The first scheme followed a two stage classification by initially identifying gender and then the emotions. The second used a divide and conquer approach, utilizing cascaded binary classifiers and the third, a parallel approach by classification with individual features, followed by a decision logic. The result of the study depicts that, these multistage classification schemes an bring improvement in the performance of speech emotion recognition system compared to the one with single stage classification. Comparable results were obtained for same experiments carried out using Emo-DB database.
Cite this Research Publication : Poorna S. S. and Nair, G. J., “Multistage Classification Scheme to Enhance Speech Emotion Recognition”, International Journal of Speech Technology, vol. 22, pp. 327–340, 2019