If a large-state Ergodic HMM is to be calculated, a large amount of memory is required. Therefore, we propose new techniques which reduce the memory requirements and computational costs significantly when the Baum-Welch algorithm is to be used. These techniques are divided into two main parts.
If is less than a given small value, this probability is set to 0.0 and there are no calculation nor memory needs for the Baum-Welch algorithm.
To put it concretely, the forward value must be
calculated as follows[1] with the basic Baum-Welch algorithm.
Instead of this equation, the following equations are used:
In experiments, we set to .
The output probability has the largest memory requirements among the parameters in the Ergodic HMM. So, in using this algorithm, the memory requirements are decreased for word output probabilities. The computational costs are also decreased at the same ratio.
To calculate a large-state Ergodic HMM, a large amount of memory is required. So, we start with a 1-state Ergodic HMM and increase the number of states successively.
If the parameters of the -state Ergodic HMM have already been estimated, the initial parameters of a -state Ergodic HMM are constructed as follows;
.
.
The symbol indicated that the integer has been rounded-up.
The output probability B is a little different, as random weights are used.
.
is normalized as
.
The training of this large-state Ergodic HMM is carried out as follows.
These sequence are repeated.
A similar technique where small probabilities are set to 0.0 is described in [5]. As such, this study's focus is mainly to reduce the computational costs and not to reduce the memory requirements. Also, as before, the number of states is kept small and there are no outstanding results.