The argument crit
sets the convergence criterion to either the
relative change in the log-likelihood or the absolute change in the
log-likelihood. The relative likelihood criterion (the default) assumes
convergence on iteration \(i\) when
\(\frac{\log L(i) - \log L(i-1)}{\log L(i-1)} < tol\).
The absolute likelihood criterion assumes convergence on iteration
\(i\) when \(\log L(i) - \log L(i-1) < tol\).
Use crit="absolute"
to invoke the latter
convergence criterion. Note that in that case, optimal values of the
tolerance parameter tol
scale with the value of the log-likelihood
(and these are not changed automagically).
Argument random.start
This is used for a (limited) random
initialization of the parameters. In particular, if
random.start=TRUE
, the (posterior) state probabilities are
randomized at iteration 0 (using a uniform distribution), i.e. the
\(\gamma\) variables (Rabiner, 1989) are sampled from the Dirichlet
distribution with a (currently fixed) value of
\(\alpha=0.1\); this results in values for each row of \(\gamma\)
that are quite close to zero and one; note that when these values are
chosen at zero and one, the initialization is similar to that used in
kmeans
. Random initialization is useful when no initial parameters can be
given to distinguish between the states. It is also useful for repeated
estimation from different starting values.
Argument classification
is used to choose either soft (default) or
hard classification of observations to states. When using soft classification, observations
are assigned to states with a weight equal to the posterior probability of
the state. When using hard classification, observations are assigned to states
according to the maximum a posteriori (MAP) states (i.e., each observation
is assigned to one state, which is determined by the Viterbi algorithm in the
case of depmix
models). As a result, the EM algorithm will find a local
maximum of the classification likelihood (Celeux & Govaert, 1992).
Warning: hard classification is an experimental feature,
especially for hidden Markov models, and its use is currently not advised.