GE
4/15/2015
Virtual Online Input Capable Expression Dissector
An online application that can predict future word output based of previous inputs.
Useful for:
This application was developed by learning from actual human authors. At it's heart is a what I call a stupid N-gram back-off model. Millions of internet posts were consumed from:
Features were extracted from this data, cleaned, and stored as N-grams. An N-gram is a series of words found in a certain order in text. As an example “New York City” is a frequently found 3-gram, also called a trigram.
The application looks at the last 3 words, and sees if it can find a matching 4-gram in it's database. If it does not, it searches for a matching trigram, then bigram, and ultimately unigram. It defaults to the longest and most popular gram.
The same procedure is used for shorter word inputs, but searching through smaller N-grams first.
Each predicted word takes less than a second to find. If multiple words are requested at once, then each word is appended to the end of the input text, and the model is run again.
Since this application can be trained with any human text input, we could use it to generate interesting and often funny bots. Some ideas:
However you use it, be careful to sanitize your input or you'll end up repeating Microsoft's mistake of creating an evil robot.