We investigate multimodal interface that utilizes non-verbal information like gaze directions or gestures in collaborative work in a multi-display environment. A system which allows users to manipulate windows and cursors using their heads' positions and directions are implemented. An experiment and a questionnaire are conducted. As a result, we found that the system improved the tasks in aspects of both efficiency and user's preference.