Acoustic feature selection for automatic emotion recognition from speech

作者:

Highlights:

摘要

Emotional expression and understanding are normal instincts of human beings, but automatical emotion recognition from speech without referring any language or linguistic information remains an unclosed problem. The limited size of existing emotional data samples, and the relative higher dimensionality have outstripped many dimensionality reduction and feature selection algorithms. This paper focuses on the data preprocessing techniques which aim to extract the most effective acoustic features to improve the performance of the emotion recognition. A novel algorithm is presented in this paper, which can be applied on a small sized data set with a high number of features. The presented algorithm integrates the advantages from a decision tree method and the random forest ensemble. Experiment results on a series of Chinese emotional speech data sets indicate that the presented algorithm can achieve improved results on emotional recognition, and outperform the commonly used Principle Component Analysis (PCA)/Multi-Dimensional Scaling (MDS) methods, and the more recently developed ISOMap dimensionality reduction method.

论文关键词:43.72.Ne,43.71.Bp,43.71.Ft,Emotion recognition,Feature selection,Machine learning

论文评审过程:Received 30 May 2008, Revised 18 September 2008, Accepted 18 September 2008, Available online 31 October 2008.

论文官网地址:https://doi.org/10.1016/j.ipm.2008.09.003