ID: cmp-lg/9708013

explanation-based learning of data oriented parsing

August 20, 1997

View on ArXiv

Similar papers 5

Speeding Up Natural Language Parsing by Reusing Partial Results

April 6, 2019

84% Match
Michalina Strzyz, Carlos Gómez-Rodríguez
Computation and Language

This paper proposes a novel technique that applies case-based reasoning in order to generate templates for reusable parse tree fragments, based on PoS tags of bigrams and trigrams that demonstrate low variability in their syntactic analyses from prior data. The aim of this approach is to improve the speed of dependency parsers by avoiding redundant calculations. This can be resolved by applying the predefined templates that capture results of previous syntactic analyses and d...

Find SimilarView on arXiv

Apportioning Development Effort in a Probabilistic LR Parsing System through Evaluation

April 12, 1996

84% Match
John University of Sussex Carroll, Ted University of Cambridge Briscoe
Computation and Language

We describe an implemented system for robust domain-independent syntactic parsing of English, using a unification-based grammar of part-of-speech and punctuation labels coupled with a probabilistic LR parser. We present evaluations of the system's performance along several different dimensions; these enable us to assess the contribution that each individual part is making to the success of the system as a whole, and thus prioritise the effort to be devoted to its further enha...

Find SimilarView on arXiv

Bagging and Boosting a Treebank Parser

June 5, 2000

84% Match
John C. Henderson, Eric Brill
Computation and Language

Bagging and boosting, two effective machine learning techniques, are applied to natural language parsing. Experiments using these techniques with a trainable statistical parser are described. The best resulting system provides roughly as large of a gain in F-measure as doubling the corpus size. Error analysis of the result of the boosting technique reveals some inconsistent annotations in the Penn Treebank, suggesting a semi-automatic method for finding inconsistent treebank ...

Find SimilarView on arXiv

Precision-biased Parsing and High-Quality Parse Selection

May 20, 2012

84% Match
Yoav Goldberg, Michael Elhadad
Computation and Language

We introduce precision-biased parsing: a parsing task which favors precision over recall by allowing the parser to abstain from decisions deemed uncertain. We focus on dependency-parsing and present an ensemble method which is capable of assigning parents to 84% of the text tokens while being over 96% accurate on these tokens. We use the precision-biased parsing task to solve the related high-quality parse-selection task: finding a subset of high-quality (accurate) trees in a...

Find SimilarView on arXiv

Notes About a More Aware Dependency Parser

July 20, 2015

84% Match
Matteo Grella
Computation and Language

In this paper I explain the reasons that led me to research and conceive a novel technology for dependency parsing, mixing together the strengths of data-driven transition-based and constraint-based approaches. In particular I highlight the problem to infer the reliability of the results of a data-driven transition-based parser, which is extremely important for high-level processes that expect to use correct parsing results. I then briefly introduce a number of notes about a ...

Find SimilarView on arXiv

Incorporating Semi-supervised Features into Discontinuous Easy-First Constituent Parsing

September 12, 2014

84% Match
Yannick Versley
Computation and Language

This paper describes adaptations for EaFi, a parser for easy-first parsing of discontinuous constituents, to adapt it to multiple languages as well as make use of the unlabeled data that was provided as part of the SPMRL shared task 2014.

Find SimilarView on arXiv

Memory-Based Shallow Parsing

April 24, 2002

84% Match
Erik F. Tjong Kim Sang
Computation and Language

We present memory-based learning approaches to shallow parsing and apply these to five tasks: base noun phrase identification, arbitrary base phrase recognition, clause detection, noun phrase parsing and full parsing. We use feature selection techniques and system combination methods for improving the performance of the memory-based learner. Our approach is evaluated on standard data sets and the results are compared with that of other systems. This reveals that our approach ...

Find SimilarView on arXiv

An Empirical Evaluation of Probabilistic Lexicalized Tree Insertion Grammars

August 4, 1998

84% Match
Rebecca Harvard University Hwa
Computation and Language

We present an empirical study of the applicability of Probabilistic Lexicalized Tree Insertion Grammars (PLTIG), a lexicalized counterpart to Probabilistic Context-Free Grammars (PCFG), to problems in stochastic natural-language processing. Comparing the performance of PLTIGs with non-hierarchical N-gram models and PCFGs, we show that PLTIG combines the best aspects of both, with language modeling capability comparable to N-grams, and improved parsing performance over its non...

Find SimilarView on arXiv

Learning Executable Semantic Parsers for Natural Language Understanding

March 22, 2016

84% Match
Percy Liang
Computation and Language
Artificial Intelligence

For building question answering systems and natural language interfaces, semantic parsing has emerged as an important and powerful paradigm. Semantic parsers map natural language into logical forms, the classic representation for many important linguistic phenomena. The modern twist is that we are interested in learning semantic parsers from data, which introduces a new layer of statistical and computational issues. This article lays out the components of a statistical semant...

Find SimilarView on arXiv

Cascaded Markov Models

June 6, 1999

84% Match
Thorsten Brants
Computation and Language

This paper presents a new approach to partial parsing of context-free structures. The approach is based on Markov Models. Each layer of the resulting structure is represented by its own Markov Model, and output of a lower layer is passed as input to the next higher layer. An empirical evaluation of the method yields very good results for NP/PP chunking of German newspaper texts.

Find SimilarView on arXiv