Factored LMs and interpolated models
Tanel Alumäe
tanel.alumae at aqris.com
Thu May 6 23:52:35 PDT 2004
> Let me know if this helps or if I have misunderstood your question...
>
Hello,
First, thanks to everybody for help.
My goal was, as Katrin correctly assumed, "to interpolate a
traditional class-based model and a standard n-gram model but you want
to express this within a single FLM file". This is currently not
possible, but it's not very important because I learned that I can
use:
ngram -factored -lm <FLM1> -mix-lm <FLM2>
The above really works.
Still, I noticed a strange thing with perplexity calculation. Namely,
the perplexity figures calculated by fngram and ngram are slightly
different. I used the following options and got following results:
fngram -ppl <testtext> -factor-file tmp/fngram_m.conf
Result:
61 sentences, 1009 words, 26 OOVs
0 zeroprobs, logprob= -2760.87 ppl= 441.076 ppl1= 643.604
ngram -factored -ppl <testtext> -lm tmp/fngram_m.conf 61 sentences, 1009
words,
Result:
26 OOVs 0 zeroprobs, logprob= -2761.16 ppl= 441.359 ppl1= 644.042
--
The above is for a FLM that in fact is standard word trigram. The
difference is very small.
However, when I test a FLM that is a word-given-two-previous-classes
trigram, the difference is much larger:
fngram -ppl <testtext> -factor-file tmp/fngram_c.conf
61 sentences, 1009 words, 26 OOVs
0 zeroprobs, logprob= - 2826.73 ppl= 510.034 ppl1= 750.963
And the same with ngram:
ngram -factored -lm tmp/fngram_c.conf -ppl <testtext>
61 sentences, 1009 words, 26 OOVs
0 zeroprobs, logprob= -2863.71 ppl= 553.378 ppl1= 818.917
As you see, here the difference (ppl1= 750 vs 818) is significant. Could
this be a configuration issue, a bug or have I understood smth wrong?
Regards,
Tanel Alumäe
More information about the SRILM-User
mailing list