2013 Volume 15 Issue 4 Pages 363-374
We have developed a novel method controlling robot motions, such as gestures and utterances during a dialogue, based on the listener's understanding and interest. The understanding and interest can be recognized from back-channels and head gestures. "Back-channels" are defined as sounds like 'un' uttered by a listener during a dialogue, and "head gestures" are defined as nod and tilt motions of the listener's head. The back-channels are recognized using sound features such as power and fundamental frequency. The head gestures are recognized using the movement of the skin-color area and the optical flow data. Based on the estimated understanding and interest of the listener, the speed and size of robot motions are changed. This method was implemented in a humanoid robot called SIG2. Experiments with eight participants demonstrated that the proposed method enabled the robot to increase the listener's level of interest against the dialogue.