Statistical Natural Language Parsing: Reliable Models of Language?
Sandiway Fong, SM '86, PhD '91, Associate Professor, University of Arizona
Description: The statistical natural language linguist owes much to the University of Pennsylvania's famous Treebank project. But this giant corpus of one million words _ actually, 49 thousand sentences from the Wall Street Journal all carefully labeled for their syntactic and semantic components -- is actually both a "blessing and a curse," says Sandiway Fong. This "gold standard" list of parsed sentences, the result of more than a decade of work, has become "the only game in town,"according to Fong. Linguists developing natural language algorithms often rely on the complex Penn Treebank to construct and train probabilistic, context"free grammars, and Fong acknowledges the Treebank's revolutionary impact on the field. But he also thinks it' sworthwhile to examine how systems that rely on Penn Treebank actually perform. He has been exploring three basic questions: Do such systems attain cognitively plausible knowledge of language, such as distinguishing between grammatical and ungrammatical components of sentences? How brittle are these systems, so that if you misspell a word or flip one part of the sentence, the system will "give you back some parse? Can these systems learn non"natural languages? Fong has unearthed some interesting issues. For instance, two well"known parsing systems couldn't score more than 50% figuring out the right way to pronounce the word "read" in eight sentences that deployed the past and present tenses (e.g., The girls will read the paper; The girls have read the paper). And the two systems didn't get the same sentences wrong. Fong wonders if "reading the Wall Street Journal is not a good way to learn how to pronounce 'read' or 'red.'" Fong also demonstrated that a parsing system could be turned on the presence (or absence) of a single example involving the phrase "milk with 4% butterfat," calling in question whether such systems are truly robust. While Treebank"based parsing systems demonstrably perform well on Treebank"like sentences, one cannot infer they have necessarily achieved grammatical competence nor linguistic stability. We must understand, says Fong, that 40 thousand training samples do not really provide enough parameters to provide the broad range of linguistic cases for computational systems that ordinary people pick up nearly effortlessly. "We expect statistical systems to be able to deal with noise. But they are extremely fragile, despite their statistical nature and training over a large data set."
About the Speaker(s): Sandiway Fong received his B.Sc. in Computing Science, at Imperial College of Science and Technology, University of London. He received an S.M. in 1986 at MIT, where he worked in the Artificial Intelligence Laboratory.
After working at IBM's Watson Research Center, he returned to MIT for his Ph.D.
In 1991, he joined the NEC Research Institute to work on natural language processing, and machine translation. In 2003, he moved to the University of Arizona, where his research interests are at the intersection of computer science and formal linguistics, with a focus on multilingual parsing, ontolinguistics, computational lexical semantics and computational morphology.
Host(s): School of Engineering, Laboratory for Information and Decision Systems
It looks like no one has posted a comment yet. You can be the first!
More from MIT World — special events and lectures
Added over 3 years ago | 00:52:17 | 3391 views
Added over 3 years ago | 00:53:46 | 4490 views
Added over 3 years ago | 00:39:50 | 2065 views
Added over 3 years ago | 00:48:24 | 3897 views
Added over 3 years ago | 01:36:00 | 10546 views
Added over 3 years ago | 00:53:36 | 3200 views