Discriminative training of tied-mixture HMM by deterministic annealing
نویسندگان
چکیده
A deterministic annealing algorithm for the design of tiedmixture HMM recognizers is proposed, which reduces the training sensitivity to parameter initialization, automatically smoothes the classification error cost function to allow gradientbased optimization, and seeks better solutions than known techniques. The new approach introduces randomness into the classification rule during the training process, and minimizes the expected error rate while controlling the level of randomness via a constraint on the Shannon entropy. As the entropy constraint is gradually relaxed, the effective cost function converges to the classification error rate and the system becomes a hard (nonrandom) recognizer. Experiments show that the proposed method outperforms design by maximum likelihood reestimation and by generalized probabilistic descent.
منابع مشابه
Connectionist Optimisation of Tied Mixture Hidden Markov Models
Horacio Franco Michael Cohen SRI International Menlo Park CA 94025 USA Issues relating to the estimation of hidden Markov model (HMM) local probabilities are discussed. In particular we note the isomorphism of radial basis functions (RBF) networks to tied mixture density modellingj additionally we highlight the differences between these methods arising from the different training criteria emplo...
متن کاملGMM-Free Flat Start Sequence-Discriminative DNN Training
Recently, attempts have been made to remove Gaussian mixture models (GMM) from the training process of deep neural network-based hidden Markov models (HMM/DNN). For the GMM-free training of a HMM/DNN hybrid we have to solve two problems, namely the initial alignment of the frame-level state labels and the creation of context-dependent states. Although flat-start training via iteratively realign...
متن کاملA Tied-Mixture 2-D HMM Face Recognition System
In this paper, a simplified 2-D second-order Hidden Markov Model (HMM) with tied state mixtures is applied to the face recognition problem. The mixture of the model states is fully-tied across all models for lower complexity. Tying HMM parameters is a well-known solution for the problem of insufficient training data leading to nonrobust estimation. We show that parameter tying in HMM also enhan...
متن کاملA comparison of hybrid HMM architectures using global discriminative training
This paper presents a comparison of di erent model architectures for TIMIT phoneme recognition. The baseline is a conventional diagonal covariance Gaussian mixture HMM. This system is compared to two di erent hybrid MLP/HMMs, both adhering to the same restrictions regarding input context and output states as the Gaussian mixtures. All free parameters in the three systems are jointly optimised u...
متن کاملDiscriminative optimization of large vocabulary Mandarin conversational speech recognition system
This paper examines techniques of discriminative optimization for acoustic model, including both HMM parameters and linear transforms, in the context of HUB5 Mandarin large vocabulary speech recognition task, with the aim to partly solve the problems brought by the sparseness and the highly ambiguous nature of the telephony conversational speech data. Three techniques are studied: MMI training ...
متن کامل