Title
A Neural Beamspace-Domain Filter for Real-Time Multi-Channel Speech Enhancement
Abstract
Most deep-learning-based multi-channel speech enhancement methods focus on designing a set of beamforming coefficients, to directly filter the low signal-to-noise ratio signals received by microphones, which hinders the performance of these approaches. To handle these problems, this paper designs a causal neural filter that fully exploits the spectro-temporal-spatial information in the beamspace domain. Specifically, multiple beams are designed to steer towards all directions, using a parameterized super-directive beamformer in the first stage. After that, a deep-learning-based filter is learned by, simultaneously, modeling the spectro-temporal-spatial discriminability of the speech and the interference, so as to extract the desired speech, coarsely, in the second stage. Finally, to further suppress the interference components, especially at low frequencies, a residual estimation module is adopted, to refine the output of the second stage. Experimental results demonstrate that the proposed approach outperforms many state-of-the-art (SOTA) multi-channel methods, on the generated multi-channel speech dataset based on the DNS-Challenge dataset.
Year
DOI
Venue
2022
10.3390/sym14061081
SYMMETRY-BASEL
Keywords
DocType
Volume
multi-channel speech enhancement, neural beam filter, deep learning
Journal
14
Issue
ISSN
Citations 
6
2073-8994
0
PageRank 
References 
Authors
0.34
0
7
Name
Order
Citations
PageRank
Wenzhe Liu122.07
Andong Li200.34
Xiao Wang326.24
Minmin Yuan400.34
Yi Chen55420.72
Chengshi Zheng600.34
Xiaodong Li74814.00