Title
Using vision to improve sound source separation
Abstract
We present a method of improving sound source separation using vision. The sound source separation is an essential function to accomplish auditory scene understanding by separating stream of sounds generated from multiple sound sources. By separating a stream of sounds, recognition process, such as speech recognition, can simply work on a single stream, not mixed sound of several speakers. The performance is known to be improved by using stereo/binaural microphone and microphone array which provides spatial information for separation. However, these methods still have more than 20 degree of positional ambiguities. In this paper, we further added visual information to provide more specific and accurate position information. As a result, separation capability was drastically improved. In addition, we found that the use of approximate direction information drastically improve object tracking accuracy of a simple vision system, which in turn improves performance of the auditory system. We claim that the integration of vision and auditory inputs improves performance of tasks in each perception, such as sound source separation and object tracking, by bootstrapping.
Year
Venue
Keywords
1999
AAAI/IAAI
separation capability,auditory input,accurate position information,mixed sound,spatial information,approximate direction information,auditory scene understanding,sound source separation,visual information,multiple sound source,speech recognition,vision system,object tracking
Field
DocType
ISBN
Computer vision,Auditory scene analysis,Machine vision,Computer science,Microphone array,Speech recognition,Video tracking,Artificial intelligence,Binaural recording,Perception,Microphone,Computational auditory scene analysis
Conference
0-262-51106-1
Citations 
PageRank 
References 
15
3.83
9
Authors
3
Name
Order
Citations
PageRank
Yukiko Nakagawa1409.78
Hiroshi G. Okuno22092233.19
Hiroaki Kitano33515539.37