Abstract | ||
---|---|---|
Recently, there has been a rising interest in sound recognition via acoustic sensor networks (ASNs) to support applications such as ambient assisted living or environmental habitat monitoring. With state-of-the-art sound recognition being dominated by deep-learning-based approaches, there is a high demand for labeled training data. Despite the availability of large-scale data sets such as Google's AudioSet, acquiring training data matching a certain application environment is still often a problem. In this paper we are concerned with human activity monitoring in a domestic environment using an ASN consisting of multiple nodes each providing multichannel signals. We propose a self-training based domain adaptation approach, which only requires unlabeled data from the target environment. Here, a sound recognition system trained on AudioSet, the teacher, generates pseudo labels for data from the target environment on which a student network is trained. The student can furthermore glean information about the spatial arrangement of sensors and sound sources to further improve classification performance. It is shown that the student significantly improves recognition performance over the pre-trained teacher without relying on labeled data from the environment the system is deployed in. |
Year | DOI | Venue |
---|---|---|
2021 | 10.23919/EUSIPCO54536.2021.9616009 | 29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021) |
Keywords | DocType | ISSN |
acoustic sensor network, sound recognition, scene classification, domain adaptation, self-training | Conference | 2076-1465 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Janek Ebbers | 1 | 0 | 0.34 |
Moritz Curt Keyser | 2 | 0 | 0.34 |
Reinhold Haeb-Umbach | 3 | 1487 | 211.71 |