[SRILM User List] how are the probabilities computed in ngram-count
Andreas Stolcke
stolcke at icsi.berkeley.edu
Wed Apr 11 10:00:59 PDT 2012
On 4/11/2012 5:48 AM, Saman Noorzadeh wrote:
> Thank you,
> -cdiscount 0 works perfectly, but now that I have read about smoothing
> and different methods of discounting I have another question:
>
> I want to know your ideas about this problem:
> I want to have a model out of a text. and then predict what the user
> is typing (a word prediction approach). at any moment I will predict
> what the next character would be according to my bigrams.
> Do you think methods of discounting and smoothing are useful in
> treating the training data?
> or it is more appropriate if I just disable it?
It probably won't make a difference because in an application like this
you are interested in finding the most probable next tokens, and
smoothing helps you with the least probable tokens. However, this type
of LM application has been studied extensively, and you should look
online what others have done. Try
http://scholar.google.com/scholar?q=character+prediction+typing&hl=en&btnG=Search&as_sdt=1%2C5&as_sdtp=on
Andreas
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.speech.sri.com/pipermail/srilm-user/attachments/20120411/994046cb/attachment.html>
More information about the SRILM-User
mailing list