[SRILM User List] How to train LM fast with large corpus

Meng Chen chenmengdx at gmail.com
Wed Jul 18 03:40:37 PDT 2012


Hi, I want to ask how to train N-gram language model with SRILM if the
corpus is very large (100GB). Should I still use the command of *ngram-count
*? Or use *make-big-lm* instead? I also want to know if there is any
limitation of training corpus in vocabulary and size with SRILM?
Thanks!

Meng CHEN
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.speech.sri.com/pipermail/srilm-user/attachments/20120718/2e669fa6/attachment.html>


More information about the SRILM-User mailing list