built on 4 orders of n-grams
Markov Chains assumption: probability of the next event depends only on the immediate previous event(s), not on the whole events history
Backoff strategy: tetragrams → trigrams → bigrams → unigrams
How it works
The model uses a 4-gram backoff model to predict the next word.