- Model compression during loading trough N-gram context hashing
- Memory footpring reduction of a factor of 2.82
- Selection of most probable words for each N-gram order using backoff weights and continuation probabilities
- Averaging of probabilities through all N-gram orders
- Ordering by decreasing scores and N-gram order
For a detailed description, see the Algorithm tab on the Web App.