Previous [ 1] [ 2] [ 3] [ 4] [ 5] [ 6] [ 7] [ 8] [ 9] [ 10] [ 11] [ 12] [ 13] [ 14] [ 15] [ 16] [ 17] [ 18] [ 19] [ 20] [ 21]

@

Journal of Information Science and Engineering, Vol. 31 No. 3, pp. 799-820 (May 2015)


Duration High-Order Hidden Markov Models and Training Algorithms for Speech Recognition*


LEE-MIN LEE
Department of Electrical Engineering
Da-Yeh University
Changhua, 515 Taiwan
E-mail: lmlee@mail.dyu.edu.tw

The duration high-order hidden Markov model (DHO-HMM) can capture the dynamic evolution of a physical system more precisely than can the first-order hidden Markov model (HMM). The relations among the DHO-HMM, high-order HMM (HOHMM), hidden semi-Markov model (HSMM), and HMM are presented and discussed. Recursive forward and backward probability functions for the partial observation sequence were derived, and were used to calculate the expected number of state transitions and to update the DHO-HMMs parameters. Viterbi decoding and training algorithms for the DHO-HMM are also presented. Experimental results show that the proposed expectation- maximization (EM) training algorithm can obtain more reliable and accurate estimate of DHO-HMMs than the Viterbi training method. Experimental results also show that the DHO-HMM speech recognizer is superior to the HSMM and the baseline conventional HMM recognizers. In experiments, the DHO-HMM speech recognizer trained by the EM algorithm reduces recognition errors by up to 53% compared with the baseline HMM.

Keywords: EM algorithm, hidden Markov model, high-order, speech recognition, duration modeling

Full Text () Retrieve PDF document (201505_03.pdf)

Received April 23, 2014; revised July 15, 2014; accepted August 13, 2014.
Communicated by Chung-Hsien Wu.