There has been an increase in interest in the development of human-interface system that incorporates multimodal input-methods.
In this study, an Automaton-Controlled One Pass Viterbi program was developed to recognize speaker independent continuous speech. The base-line recognition mechanism is based on the One Pass Viterbi algorithm using the concatenation of syllable HMMs(Hidden Markov Models). In this program, if the dictionary and automaton information files based on the definition of finite state automaton are prepared, all kinds of sentences for the automaton can be accepted.
As an application of the speech recognition, the input methods by speech and by touch-screen were studied. Since these two methods are complementary each other, they can be used together to create more userfriendly human interface. To illustrate this, a multi-modal robot control simulation system was built on the Sun Sparc10 (using the attached AD converter with 11.025kHz sampling rate). To evaluate this system, a robot control simulation system using only the speech input method and a simulation system using only the touch-screen input method were also developed. After evalvating these systems, it was found out that the multi-modal system incorporating both speech input and touch-screen input was able to cover up the deficiencies that were in the use of only touch-screen input and to improve speech recognition speed and error rate. Also, by incorporating both speech input and touch-screen input, the operational errors became infrequent and more kinds of useful operations could be performed as compared to using only speech or only touch-screen inputs.
抄録全体を表示