Abstract | ||
---|---|---|
Most state-of-the-art speaker recognition systems are partially or completely based on Gaussian mixture models (GMM). GMM have been widely and successfully used in speaker recognition during the last decades. They are traditionally estimated from a world model using the generative criterion of Maximum A Posteriori. In an earlier work, we proposed an efficient algorithm for discriminative learning of GMM with diagonal covariances under a large margin criterion. In this paper, we evaluate the combination of the large margin GMM modeling approach with SVM in the setting of speaker identification. We carry out a full NIST speaker identification task using NIST-SRE'2006 data, in a Symmetrical Factor Analysis compensation scheme. The results show that the two modeling approaches are complementary and that their combination outperforms their single use. |
Year | Venue | Keywords |
---|---|---|
2013 | 2013 PROCEEDINGS OF THE 21ST EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO) | Large margin training, Gaussian mixture models, discriminative learning, Support vector machines, speaker recognition |
Field | DocType | Citations |
Diagonal,Speaker identification,Pattern recognition,Computer science,Support vector machine,Speech recognition,Speaker recognition,NIST,Artificial intelligence,Maximum a posteriori estimation,Mixture model,Discriminative learning | Conference | 0 |
PageRank | References | Authors |
0.34 | 15 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Reda Jourani | 1 | 3 | 1.14 |
Khalid Daoudi | 2 | 145 | 23.68 |
Régine André-Obrecht | 3 | 122 | 19.19 |
Driss Aboutajdine | 4 | 589 | 88.82 |