TY - JOUR
T1 - Gaussian mixture models reduction by variational maximum mutual information
AU - Bar-Yosef, Yossi
AU - Bistritz, Yuval
N1 - Publisher Copyright:
© 1991-2012 IEEE.
PY - 2015/3/15
Y1 - 2015/3/15
N2 - Gaussian mixture models (GMMs) are widely used in a variety of classification tasks where it is often important to approximate high order models by models with fewer components. The paper proposes a novel approach to this problem based on a parametric realization of the maximum mutual information (MMI) criterion and its approximation by a closed-form expression named variational-MMI (VMMI). The maximization of the VMMI can be carried out in an analytically tractable manner and it aims at improving the discrimination ability of the reduced set of models, a goal that was not targeted in previous approaches that simplify each class-related GMM independently. Two effective algorithms are proposed and studied for the optimization of the VMMI criterion. One is a steepest descent type algorithm, and the other, called line search A-functions (LSAF), uses concave associated functions. Experiments held in two speech related tasks, phone recognition and language recognition, demonstrate that the VMMI-based parametric model reduction algorithms significantly outperform previous non-discriminative methods. According to these experiments, the EM-like LSAF-based algorithm requires less iterations and converges to a better value of the objective function compared to the steepest descent algorithm.
AB - Gaussian mixture models (GMMs) are widely used in a variety of classification tasks where it is often important to approximate high order models by models with fewer components. The paper proposes a novel approach to this problem based on a parametric realization of the maximum mutual information (MMI) criterion and its approximation by a closed-form expression named variational-MMI (VMMI). The maximization of the VMMI can be carried out in an analytically tractable manner and it aims at improving the discrimination ability of the reduced set of models, a goal that was not targeted in previous approaches that simplify each class-related GMM independently. Two effective algorithms are proposed and studied for the optimization of the VMMI criterion. One is a steepest descent type algorithm, and the other, called line search A-functions (LSAF), uses concave associated functions. Experiments held in two speech related tasks, phone recognition and language recognition, demonstrate that the VMMI-based parametric model reduction algorithms significantly outperform previous non-discriminative methods. According to these experiments, the EM-like LSAF-based algorithm requires less iterations and converges to a better value of the objective function compared to the steepest descent algorithm.
KW - Continuous-discrete MMI
KW - Gaussian mixture models reduction
KW - discriminative learning
KW - hierarchical clustering
UR - http://www.scopus.com/inward/record.url?scp=84923862744&partnerID=8YFLogxK
U2 - 10.1109/TSP.2015.2398844
DO - 10.1109/TSP.2015.2398844
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:84923862744
SN - 1053-587X
VL - 63
SP - 1557
EP - 1569
JO - IEEE Transactions on Signal Processing
JF - IEEE Transactions on Signal Processing
IS - 6
M1 - 7027858
ER -