Perplexity in "ngram"

Mats Svenson svmats at yahoo.com
Tue May 29 23:57:12 PDT 2007


Hi,
 I have tried to use "ngram" to count perplexity of my
LMs. However, I am not sure how does the srilm
implementation treat OOVs in terms of counting
perplexity. Is it that "log P(<unk>|history) != 0" or
OOVs are just ignored? If a model with a higher number
of OOVs has a lower perplexity than another LM, does
it mean that it is "better" in this -ppl
implementation?

Second, in some discussions, I have heard about -ppl1
option, but the current version does not seem to have
it. In what -ppl1 differs from -ppl?

Third, is there a way how to meaningfully compute
perplexity for a hidden event LM? Or another way how
to evaluate hidden event LM quality?

Thanks for your help,
 Mats


 
____________________________________________________________________________________
Get your own web address.  
Have a HUGE year through Yahoo! Small Business.
http://smallbusiness.yahoo.com/domains/?p=BESTDEAL



More information about the SRILM-User mailing list