HMM training

2 road lines:

  • EM algorithm(Baum-Welch)
  • Gradient Descent

The main advantage of the Baum-Welch algorithm (and hence the ML training) is due to its simplicity and the fact that requires no parameter tuning. Furthermore compared to standard gradient descent, even for ML training, the Baum-Welch algorithm achieves significantly faster convergence rates. On the other hand gradient descent (especially in the case of large models) requires careful search in the parameter space for an appropriate learning rate in order to achieve the best possible performance.

References:

  1. Pantelis G. Bagos, Theodore Liakopoulos, Stavros J. Hamodrakas: Faster Gradient Descent Training of Hidden Markov Models, Using Individual Learning Rate Adaptation. ICGI 2004: 40-52
  2. Baldi, P., Chauvin, Y.: Smooth On-Line Learning Algorithms for Hidden Markov Models. Neural Comput. 6(2) (1994) 305-316

你可能感兴趣的:(hmm)