Title
Two-Way Feature Extraction for Speech Emotion Recognition Using Deep Learning
Abstract
Recognizing human emotions by machines is a complex task. Deep learning models attempt to automate this process by rendering machines to exhibit learning capabilities. However, identifying human emotions from speech with good performance is still challenging. With the advent of deep learning algorithms, this problem has been addressed recently. However, most research work in the past focused on feature extraction as only one method for training. In this research, we have explored two different methods of extracting features to address effective speech emotion recognition. Initially, two-way feature extraction is proposed by utilizing super convergence to extract two sets of potential features from the speech data. For the first set of features, principal component analysis (PCA) is applied to obtain the first feature set. Thereafter, a deep neural network (DNN) with dense and dropout layers is implemented. In the second approach, mel-spectrogram images are extracted from audio files, and the 2D images are given as input to the pre-trained VGG-16 model. Extensive experiments and an in-depth comparative analysis over both the feature extraction methods with multiple algorithms and over two datasets are performed in this work. The RAVDESS dataset provided significantly better accuracy than using numeric features on a DNN.
Year
DOI
Venue
2022
10.3390/s22062378
SENSORS
Keywords
DocType
Volume
speech emotion recognition, machine learning, neural network
Journal
22
Issue
ISSN
Citations 
6
1424-8220
0
PageRank 
References 
Authors
0.34
0
8
Name
Order
Citations
PageRank
Apeksha Aggarwal101.01
Akshat Srivastava200.34
Ajay Agarwal300.34
Nidhi Chahal400.34
Dilbag Singh56715.16
Abeer Ali Alnuaim601.01
Aseel Alhadlaq700.34
Heung-No Lee805.41