Speaker Selection and Tracking in a Cluttered Environment with Audio and Visual Information

Authors
Lim, YoonseobChoi, Jongsuk
Issue Date
2009-08
Publisher
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
Citation
IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, v.55, no.3, pp.1581 - 1589
Abstract
Presented in this paper is a data association method using audio and visual data which localizes targets in a cluttered environment and detects who is speaking to a robot. A particle filter is applied to efficiently select the optimal association between the target and the measurements. State variables are composed of tat-get positions and speaking states. To update the speaking state, we first evaluate the incoming sound signal based on cross-correlation and then calculate a likelihood front the audio information. The visual measurement is used to find an optimal association between the tat-get and the observed objects. The number of targets that the robot should interact with is updated from the existence probabilities and associations.,forehand and Experimental data were collected be simulated on a computer to verify the performance of the proposed method applied to the speaker selection problem in a cluttered environment. The algorithm was also implemented tit a robotic system to demonstrate reliable interactions between the robot and speaking targets.
Keywords
Speaker Localization; Human-Robot Interaction; Data Association; Particle Filter
ISSN
0098-3063
URI
https://pubs.kist.re.kr/handle/201004/132276
DOI
10.1109/TCE.2009.5278030
Appears in Collections:
KIST Article > 2009
Files in This Item:
There are no files associated with this item.
Export
RIS (EndNote)
XLS (Excel)
XML

qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE