Training hidden Markov models with multiple observations - A combinatorial method

被引:95
作者
Li, XL
Parizeau, M
Plamondon, R
机构
[1] CADlink Technol Corp, Ottawa, ON K1H 1E1, Canada
[2] Univ Laval, Dept Genie Elect & Genie Informat, St Foy, PQ G1K 7P4, Canada
[3] Ecole Polytech, Montreal, PQ H3C 3A7, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
hidden Markov model; forward-backward procedure; Baum-Welch algorithm; multiple observation training;
D O I
10.1109/34.845379
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
Hidden Markov models (HMMs) are stochastic models capable of statistical learning and classification. They have been applied in speech recognition and handwriting recognition because of their great adaptability and Versatility in handling sequential signals. On the other hand, as these models have a complex structure and also because the involved data sets usually contain uncertainty, it is difficult to analyze the multiple observation training problem without certain assumptions. For many years researchers have used Levinson's training equations in speech and handwriting applications, simply assuming that all observations are independent of each other. This paper presents a formal treatment of HMM multiple observation training without imposing the above assumption. In this treatment, the multiple observation probability is expressed as a combination of individual observation probabilities without losing generality. This combinatorial method gives one more freedom in making different dependence-independence assumptions. By generalizing Baum's auxiliary function into this framework and building up an associated objective function using the Lagrange multiplier method, it is proven that the derived training equations guarantee the maximization of the objective function. Furthermore, we show that Levinson's training equations can be easily derived as a special case in this treatment.
引用
收藏
页码:371 / 377
页数:7
相关论文
共 23 条
[1]
[Anonymous], 1989, Automatic speech recognition: The development of the SPHINX system
[2]
A MAXIMUM-LIKELIHOOD APPROACH TO CONTINUOUS SPEECH RECOGNITION [J].
BAHL, LR ;
JELINEK, F ;
MERCER, RL .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1983, 5 (02) :179-190
[3]
SMOOTH ONLINE LEARNING ALGORITHMS FOR HIDDEN MARKOV-MODELS [J].
BALDI, P ;
CHAUVIN, Y .
NEURAL COMPUTATION, 1994, 6 (02) :307-318
[4]
GROWTH TRANSFORMATIONS FOR FUNCTIONS ON MANIFOLDS [J].
BAUM, LE ;
SELL, GR .
PACIFIC JOURNAL OF MATHEMATICS, 1968, 27 (02) :211-&
[5]
STATISTICAL INFERENCE FOR PROBABILISTIC FUNCTIONS OF FINITE STATE MARKOV CHAINS [J].
BAUM, LE ;
PETRIE, T .
ANNALS OF MATHEMATICAL STATISTICS, 1966, 37 (06) :1554-&
[6]
[7]
A MAXIMIZATION TECHNIQUE OCCURRING IN STATISTICAL ANALYSIS OF PROBABILISTIC FUNCTIONS OF MARKOV CHAINS [J].
BAUM, LE ;
PETRIE, T ;
SOULES, G ;
WEISS, N .
ANNALS OF MATHEMATICAL STATISTICS, 1970, 41 (01) :164-&
[8]
BAUM LE, 1970, INEQUALITY, V3, P1
[9]
A FAST STATISTICAL MIXTURE ALGORITHM FOR ONLINE HANDWRITING RECOGNITION [J].
BELLEGARDA, EJ ;
BELLEGARDA, JR ;
NAHAMOO, D ;
NATHAN, KS .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1994, 16 (12) :1227-1233
[10]
AN HMM/MLP ARCHITECTURE FOR SEQUENCE RECOGNITION [J].
CHO, SB ;
KIM, JH .
NEURAL COMPUTATION, 1995, 7 (02) :358-369