Mathematical Foundations of Computational Linguistics

Instructor: David McAllester

MWF 10:30 Ry 276

This course will cover various mathematical concepts relevant to computational linguistics. The course covers both statistical and logical concepts. The statistical material will include hidden Markov models (HMMs); probabilistic context free grammars; n-gram language modeling; expectation-maximization (EM); exponential models (maximum entropy methods); and various formal generalization guarantees from machine learning theory. The logical material will include Montague grammar, the simply type lambda calculus, and a type-denotational approach to tense and aspect. The relationship between statistics and logic is left as a major open problem of computational linguistics.

This page contains lecture slides in postscript format. The last slide in each lecture (other than the first) contains an assignment. For registered students the assignments from Monday, Wednesday and Friday are due the following Wednesday. The first assignment (the two problems from the second and third lecture) is due at Lecture Oct. 8.