mix LM
Andreas Stolcke
stolcke at speech.sri.com
Tue Sep 3 09:13:18 PDT 2002
In message <3D74DD88.8E7C2BD8 at inzigo.com>you wrote:
> His first suggestion reminds me the mixture LM. Actually I made some
> tests on the interpolattion approach, including class + word LM. I
> always found that the perplexity (and WER) is a linear function of the
> interpolation parameter (Lambda), so the best results are always at the
> ends, which makes the interpolation trivil. Did I miss something, or it
> is the case for some domains?
>
Hongqin,
how did you find the best interpolation weight? I hope you didn't
use trial-and-error and used the compute-best-mix script instead.
In my experience the perplexity is not a linear function of lambda,
unless maybe your class-based LM is very bad. Rather, ppl should be
U-shaped function as lambda varies between 0 and 1.
--Andreas
More information about the SRILM-User
mailing list