Title
ADAPTING SOUND RECOGNITION TO A NEW ENVIRONMENT VIA SELF-TRAINING
Abstract
Recently, there has been a rising interest in sound recognition via acoustic sensor networks (ASNs) to support applications such as ambient assisted living or environmental habitat monitoring. With state-of-the-art sound recognition being dominated by deep-learning-based approaches, there is a high demand for labeled training data. Despite the availability of large-scale data sets such as Google's AudioSet, acquiring training data matching a certain application environment is still often a problem. In this paper we are concerned with human activity monitoring in a domestic environment using an ASN consisting of multiple nodes each providing multichannel signals. We propose a self-training based domain adaptation approach, which only requires unlabeled data from the target environment. Here, a sound recognition system trained on AudioSet, the teacher, generates pseudo labels for data from the target environment on which a student network is trained. The student can furthermore glean information about the spatial arrangement of sensors and sound sources to further improve classification performance. It is shown that the student significantly improves recognition performance over the pre-trained teacher without relying on labeled data from the environment the system is deployed in.
Year
DOI
Venue
2021
10.23919/EUSIPCO54536.2021.9616009
29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021)
Keywords
DocType
ISSN
acoustic sensor network, sound recognition, scene classification, domain adaptation, self-training
Conference
2076-1465
Citations 
PageRank 
References 
0
0.34
0
Authors
3
Name
Order
Citations
PageRank
Janek Ebbers100.34
Moritz Curt Keyser200.34
Reinhold Haeb-Umbach31487211.71