Alexey Nepomnyashchiy
August 23, 2015
Word prediction is the one of most common problems in the Natural language processing.
Two smoothing methods: absolute discounting and interpolated modified Kneyser-Ney (current state of the art).
Probabilities was calculated for [1:5]-grams, but while exporting the data model, [3:5]-grams that was found in corpora only once were filtered out.
Only most frequent combinations was exported, so only one word can be suggested.
Words that weren’t found in unigrams substituted with UNK.