Time permitting, we will adjust our model in a more complex formulation to account for a message subject and a message recipient set. We conjecture that both subject and message-recipient set have a telling influence on the next-word estimate--in practice, people tend to discuss certain topics with particular groups of people.
In another step of complexity, we will adjust the model to respond to individually typed letters. Incorrect predictions will lead to user typing, and typed letters in combination with history, can provide useful evidence about the correct word.
The tutorial on hidden Markov models [4] and
chapter 9 of the NLP textbook [2] can apply
in these more complex problem formulations. The ``Statistical
Language Modeling Toolkit'' might provide useful NLP facilities for
efficiently constructing
-grams [1].