Goal:
Build a next-word prediction model and deploy it as a Shiny web application.
Training Data: - Sample text sentences
- Converted to lowercase and tokenized
Model: - Bigram (2-gram) frequency table
- Predicts next word based on last word of input
Fallback: - If no match found → returns “no_prediction”