Segment-Quantization-based HMMs (SQ-HMMs), which are trained with Segment-Quantized codebook (SQ-codebook) using the following two approaches, show higher recognition performance than VQ-HMMS in a noisy environment : 1) phoneme category dependent SQ-codebook generation from categorized speech data using phoneme label information, 2) speaking-style dependent SQ-codebook generation from the task close to input speech or the same task as input speech. In a comparison of SQ with our VQ-based baseline method in 18 consonant recognition experiments, recognition rates for SQ-HMMs are improved 3.9%, 8.2% and 9.1% at SNR=∞. 30dB and 20dB, respectively. In Japanese phrase recognition experiments, phrase recognition rates are 88.2%, 84.2% and 52.7% at SNR=∞, 30dB and 20dB, respectively. In comparison with our VQ-based baseline method, these recognition rates are improved 0.7%, 10.0% and 11.5%.
View full abstract