The impact of computer systems that can understand natural
language will be tremendous. To develop this capability we need to
be able to automatically and efficiently analyze large amounts of
text. Manually devised rules are not sufficient to provide coverage
to handle the complex structure of natural language, necessitating
systems that can automatically learn from examples. To handle the
flexibility of natural language, it has become standard practice to
use statistical models, which assign probabilities for example to
the different meanings of a word or the plausibility of grammatical
constructions.
This book develops a general coarse-to-fine framework for
learning and inference in large statistical models for natural
language processing.
Coarse-to-fine approaches exploit a sequence of models which
introduce complexity gradually. At the top of the sequence is a
trivial model in which learning and inference are both cheap. Each
subsequent model refines the previous one, until a final,
full-complexity model is reached. Applications of this framework to
syntactic parsing, speech recognition and machine translation are
presented, demonstrating the effectiveness of the approach in terms
of accuracy and speed. The book is intended for students and
researchers interested in statistical approaches to Natural
Language Processing.
"Slav s work"Coarse-to-Fine Natural Language Processing
"represents a major advance in the area of syntactic parsing, and a
great advertisement for the superiority of the machine-learning
approach."
Eugene Charniak (Brown University)"
General
Is the information for this product incomplete, wrong or inappropriate?
Let us know about it.
Does this product have an incorrect or missing image?
Send us a new image.
Is this product missing categories?
Add more categories.
Review This Product
No reviews yet - be the first to create one!