Overview of the model algorithm
- This presentation briefly explains an NLP model that predicts the next word to followed by a phrase.
- This presentation will:
- Explain the core algorithm of the model.
- Summarize the performance of the prediction algorith.
- Explain how to use the Shiny application that runs the model.
Katz’s back off model is a generative n-gram language model that estimates the conditional probability of a word given its history in the n-gram. It accomplishes this estimation by “backing-off” to models with smaller histories under certain conditions. By doing so, the model with the most reliable information about a given history is used to provide the better results.