Title
A Comparative Re-Assessment of Feature Extractors for Deep Speaker Embeddings
Abstract
Modern automatic speaker verification relies largely on deep neural networks (DNNs) trained on mel-frequency cepstral coefficient (MFCC) features. While there are alternative feature extraction methods based on phase, prosody and long-term temporal operations, they have not been extensively studied with DNN-based methods. We aim to fill this gap by providing extensive re-assessment of 14 feature extractors on VoxCeleb and SITW datasets. Our findings reveal that features equipped with techniques such as spectral centroids, group delay function, and integrated noise suppression provide promising alternatives to MFCCs for deep speaker embeddings extraction. Experimental results demonstrate up to 16.3% (VoxCeleb) and 25.1% (SITW) relative decrease in equal error rate (EER) to the baseline.
Year
DOI
Venue
2020
10.21437/Interspeech.2020-1765
INTERSPEECH
DocType
Citations 
PageRank 
Conference
1
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Xuechen Liu110.34
Md. Sahidullah232624.99
Tomi Kinnunen3132386.67