TY - GEN
T1 - Sound sources selection system by using onomatopoeic querries from multiple sound sources
AU - Yamamura, Yusuke
AU - Takahashi, Toru
AU - Ogata, Tetsuya
AU - Okuno, Hiroshi G.
PY - 2012
Y1 - 2012
N2 - Our motivation is to develop a robot that treats auditory information in real environment because auditory information is useful for animated communications or understanding our surroundings. Interactions by using sound information need an aquisition of it and a proper sound source reference between a user and a robot leads to it. Such sound source reference is difficult due to multiple sound sources generating in real environemnt, and we use onomatopoeic representations as a representation for the reference. This paper shows a system that selects a sound source specified by a user from multiple sound sources. Users use onomatopoeias in the specification, and our system separates a mixed sound and converts separated sounds into onomatopoeias for the selection. Onomatopoeais have the ambiguity that each user gives each expression to a certain sound and we create an original similarity based on Minimum Edit Distance and acoustic features for solving its problem. In experiments, our system receives a mixed sound consisting of three sounds and a user's query as inputs, and checks a count of a consistency of a sound source selected by a system and a sound source specified by a user in 100 tests. The result shows
AB - Our motivation is to develop a robot that treats auditory information in real environment because auditory information is useful for animated communications or understanding our surroundings. Interactions by using sound information need an aquisition of it and a proper sound source reference between a user and a robot leads to it. Such sound source reference is difficult due to multiple sound sources generating in real environemnt, and we use onomatopoeic representations as a representation for the reference. This paper shows a system that selects a sound source specified by a user from multiple sound sources. Users use onomatopoeias in the specification, and our system separates a mixed sound and converts separated sounds into onomatopoeias for the selection. Onomatopoeais have the ambiguity that each user gives each expression to a certain sound and we create an original similarity based on Minimum Edit Distance and acoustic features for solving its problem. In experiments, our system receives a mixed sound consisting of three sounds and a user's query as inputs, and checks a count of a consistency of a sound source selected by a system and a sound source specified by a user in 100 tests. The result shows
UR - http://www.scopus.com/inward/record.url?scp=84872362429&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84872362429&partnerID=8YFLogxK
U2 - 10.1109/IROS.2012.6385765
DO - 10.1109/IROS.2012.6385765
M3 - Conference contribution
AN - SCOPUS:84872362429
SN - 9781467317375
T3 - IEEE International Conference on Intelligent Robots and Systems
SP - 2364
EP - 2369
BT - 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2012
T2 - 25th IEEE/RSJ International Conference on Robotics and Intelligent Systems, IROS 2012
Y2 - 7 October 2012 through 12 October 2012
ER -