Abstract
Human behaviors are important information in conversation scenes. They are called non-verbal information. Especially the motions of a head like "nod", "tilt", and so on, have significant meanings as non-verbal information. Perspective n-Point problem is solved between 2D coordinates of a video sequence and 3D coordinates of a face model. Then, we can get rotation and translation parameters of the head motion. These parameters are used to quantify the motion. We performed an experiment of calculating them for 1798 frames and found some of parameters that can recognize a motion.