主催: The Japanese Society for Artificial Intelligence
会議名: 2010年度人工知能学会全国大会(第24回)
回次: 24
開催地: 長崎県長崎市 長崎ブリックホール
開催日: 2010/06/09 - 2010/06/11
In this paper, we propose a novel method for a robot to detect robot-directed speech from other speech: to distinguish speech that users speak to a robot from speech that users speak to other people or to himself. The originality of this work is the introduction of Multimodal Semantic Confidence measure, which is used for domain classification of input speech based on deciding whether the speech can be interpreted as a feasible action under the current physical situation in an object manipulation task. This measure is calculated by integrating speech, object, and motion confidence with weightings that are optimized by logistic regression. Then we integrate this measure with gaze tracking, and conduct experiments under conditions of natural human-robot interactions. Experimental results show that the proposed method achieved a high performance of 94% and 96% in average recall and precision rates for robot-directed speech detection.