TY - JOUR
T1 - Support vector machine training for improved hidden Markov modeling
AU - Sloin, Alba
AU - Burshtein, David
N1 - Funding Information:
Manuscript received September 3, 2006; revised May 11, 2007. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Ilya Pollak. This work was presented in part at the 24th IEEE Conference of Electrical And Electronics Engineers, Eilat, Israel, November 15–17, 2006. This work was supported in part by the KITE Consortium of the Israeli Ministry of Industry and Trade, by Muscle, a European network of excellence funded by the EC 6th framework IST programme, and by a fellowship from The Yitzhak and Chaya Weinstein Research Institute for Signal Processing at Tel-Aviv University.
PY - 2008/1
Y1 - 2008/1
N2 - We present a discriminative training algorithm, that uses support vector machines (SVMs), to improve the classification of discrete and continuous output probability hidden Markov models (HMMs). The algorithm uses a set of maximum-likelihood (ML) trained HMM models as a baseline system, and an SVM training scheme to rescore the results of the baseline HMMs. It turns out that the rescoring model can be represented as an unnormalized HMM. We describe two algorithms for training the unnormalized HMM models for both the discrete and continuous cases. One of the algorithms results in a single set of unnormalized HMMs that can be used in the standard recognition procedure (the Viterbi recognizer), as if they were plain HMMs. We use a toy problem and an isolated noisy digit recognition task to compare our new method to standard ML training. Our experiments show that SVM rescoring of hidden Markov models typically reduces the error rate significantly compared to standard ML training.
AB - We present a discriminative training algorithm, that uses support vector machines (SVMs), to improve the classification of discrete and continuous output probability hidden Markov models (HMMs). The algorithm uses a set of maximum-likelihood (ML) trained HMM models as a baseline system, and an SVM training scheme to rescore the results of the baseline HMMs. It turns out that the rescoring model can be represented as an unnormalized HMM. We describe two algorithms for training the unnormalized HMM models for both the discrete and continuous cases. One of the algorithms results in a single set of unnormalized HMMs that can be used in the standard recognition procedure (the Viterbi recognizer), as if they were plain HMMs. We use a toy problem and an isolated noisy digit recognition task to compare our new method to standard ML training. Our experiments show that SVM rescoring of hidden Markov models typically reduces the error rate significantly compared to standard ML training.
KW - Discriminative training
KW - Hidden Markov model (HMM)
KW - Speech recognition
KW - Support vector machine (SVM)
UR - http://www.scopus.com/inward/record.url?scp=37749027001&partnerID=8YFLogxK
U2 - 10.1109/TSP.2007.906741
DO - 10.1109/TSP.2007.906741
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:37749027001
SN - 1053-587X
VL - 56
SP - 172
EP - 188
JO - IEEE Transactions on Signal Processing
JF - IEEE Transactions on Signal Processing
IS - 1
ER -