Abstract
We present a novel generative model for natural language tree structures in which semantic (lexical dependency) and syntactic (PCFG) structures are scored with separate models. This factorization provides conceptual simplicity, straightforward opportunities for separately improving the component models, and a level of performance comparable to similar, non-factored models. Most importantly, unlike other modern parsing models, the factored model admits an extremely effective A* parsing algorithm, which enables efficient, exact inference.
Keywords
Affiliated Institutions
Related Publications
Accurate unlexicalized parsing
We demonstrate that an unlexicalized PCFG can parse much more accurately than previously shown, by making use of simple, linguistically motivated state splits, which break down ...
Morphology-based and sub-word language modeling for Turkish speech recognition
We explore morphology-based and sub-word language modeling approaches proposed for morphologically rich languages, and evaluate and contrast them for Turkish broadcast news tran...
ASPECTS OF THE THEORY OF SYNTAX
Abstract : Contents: Methodological preliminaries: Generative grammars as theories of linguistic competence; theory of performance; organization of a generative grammar; justifi...
Compilers: Principles, Techniques, and Tools
1 Introduction 1.1 Language Processors 1.2 The Structure of a Compiler 1.3 The Evolution of Programming Languages 1.4 The Science of Building a Compiler 1.5 Applications of Comp...
Improved Transition-based Parsing by Modeling Characters instead of Words with LSTMs
We present extensions to a continuousstate dependency parsing method that makes it applicable to morphologically rich languages. Starting with a highperformance transition-based...
Publication Info
- Year
- 2002
- Type
- article
- Volume
- 15
- Pages
- 3-10
- Citations
- 745
- Access
- Closed