Abstract | ||
---|---|---|
Recently, the hierarchical extreme learning machine (HELM) model has been utilized for speech enhancement (SE) and demonstrated promising performance, especially when the amount of training data is limited and the system does not support heavy computations. Based on the success of audio-only-based systems, termed AHELM, we propose a novel audio-visual HELM-based SE system, termed AVHELM that integrates the audio and visual information to confrontate the unseen non-stationery noise problem at low SNR levels to attain improved SE performance. The experimental results demonstrate that AVHELM can yield satisfactory enhancement performance with a limited amount of training data and outperforms AHELM in terms of three standardized objective measures under matched and mismatched testing conditions, confirming the effectiveness of incorporating visual information into the HELM-based SE system. |
Year | DOI | Venue |
---|---|---|
2019 | 10.23919/EUSIPCO.2019.8903105 | 2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO) |
Keywords | Field | DocType |
Speech Enhancement, Hierarchical Extreme Learning Machine, Audio-Visual, Multi-Modal | Training set,Speech enhancement,Extreme learning machine,Computer science,Speech recognition,Computation | Conference |
ISSN | Citations | PageRank |
2076-1465 | 0 | 0.34 |
References | Authors | |
0 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Tassadaq Hussain | 1 | 48 | 9.41 |
Yu Tsao | 2 | 208 | 50.09 |
Hsin-Min Wang | 3 | 0 | 1.35 |
Jia-Ching Wang | 4 | 515 | 58.13 |
Sabato Marco Siniscalchi | 5 | 310 | 30.21 |
Wen-Hung Liao | 6 | 1 | 1.03 |