Earlier in the course we introduced the idea of Principles of Inductive Inference and discussed two such principles: the Principle of Maximum Likelihood and Inference Using Conditionalization. Recall that a principle of inductive inference provides a way of choosing amongst elements of a hypothesis space in light of some evidence.

Hidden Markov models and finite-state automata introduced a new kind of hidden or latent structure into our modeling idiom: sequences of categories or states. Given a sequence of words, we must perform some kkind of inference to reason about the sequence of unobserved categories over those words. Computation done to perform such inference about the latent structure underlying sentences is so important in computational linguistics that it has a special name: parsing. Over the next several lectures, we will consider different approaches to parsing for finite-state/hidden-Markov models. These approaches can again be categorized into maximum likelihood methods which, as before, seek a single sequence of categories which maximizes the probability of the data and Bayesian methods which, as before, seek tocapture the whole conditional distribution over category sequences given the observed sequence of words.


29 Determinism and Non-Determinism 31 The Viterbi Algorithm