#8168The method combined thelog-likelihood under a baseline model (that of Collins [1999]) with evidence from an additional 500,000 features over parse trees that were not included in the original model.
model,25-11-J05-1003,ak
feature selection methods
</term>
within
<term>
log-linear ( maximum-entropy ) models
</term>
. Although the experiments in this
#8295We argue that the method is an appealing alternative—in terms of both simplicity and efficiency—to work on feature selection methods withinlog-linear ( maximum-entropy ) models.
tech,39-12-J05-1003,ak
,
<term>
speech recognition
</term>
,
<term>
machine translation
</term>
, or
<term>
natural language generation
#8340Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example, speech recognition,machine translation, or natural language generation.
model,2-3-J05-1003,ak
these
<term>
parses
</term>
. A second
<term>
model
</term>
then attempts to improve upon this
#8056A secondmodel then attempts to improve upon this initial ranking, using additional features of the tree as evidence.
model,34-7-J05-1003,ak
were not included in the original
<term>
model
</term>
. The new
<term>
model
</term>
achieved
#8198The method combined the log-likelihood under a baseline model (that of Collins [1999]) with evidence from an additional 500,000 features over parse trees that were not included in the originalmodel.
model,2-8-J05-1003,ak
original
<term>
model
</term>
. The new
<term>
model
</term>
achieved 89.75 %
<term>
F-measure
</term>
#8202The newmodel achieved 89.75% F-measure, a 13% relative decrease in F-measure error over the baseline model’s score of 88.2%.
tech,43-12-J05-1003,ak
<term>
machine translation
</term>
, or
<term>
natural language generation
</term>
. We present a novel method for discovering
#8344Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example, speech recognition, machine translation, ornatural language generation.
tech,8-12-J05-1003,ak
experiments in this article are on
<term>
natural language parsing ( NLP )
</term>
, the approach should be applicable
#8309Although the experiments in this article are onnatural language parsing ( NLP ), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example, speech recognition, machine translation, or natural language generation.
tech,23-12-J05-1003,ak
should be applicable to many other
<term>
NLP problems
</term>
which are naturally framed as
<term>
#8324Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many otherNLP problems which are naturally framed as ranking tasks, for example, speech recognition, machine translation, or natural language generation.
other,25-7-J05-1003,ak
additional 500,000
<term>
features
</term>
over
<term>
parse trees
</term>
that were not included in the original
#8189The method combined the log-likelihood under a baseline model (that of Collins [1999]) with evidence from an additional 500,000 features overparse trees that were not included in the original model.
other,24-2-J05-1003,ak
initial
<term>
ranking
</term>
of these
<term>
parses
</term>
. A second
<term>
model
</term>
then
#8052The base parser produces a set of candidate parses for each input sentence, with associated probabilities that define an initial ranking of theseparses.
other,23-9-J05-1003,ak
the
<term>
feature space
</term>
in the
<term>
parsing data
</term>
. Experiments show significant efficiency
#8249The article also introduces a new algorithm for the boosting approach which takes advantage of the sparsity of the feature space in theparsing data.
tech,11-1-J05-1003,ak
which rerank the output of an existing
<term>
probabilistic parser
</term>
. The
<term>
base parser
</term>
produces
#8025This article considers approaches which rerank the output of an existingprobabilistic parser.
other,16-2-J05-1003,ak
input sentence
</term>
, with associated
<term>
probabilities
</term>
that define an initial
<term>
ranking
#8044The base parser produces a set of candidate parses for each input sentence, with associatedprobabilities that define an initial ranking of these parses.
other,21-2-J05-1003,ak
probabilities
</term>
that define an initial
<term>
ranking
</term>
of these
<term>
parses
</term>
. A second
#8049The base parser produces a set of candidate parses for each input sentence, with associated probabilities that define an initialranking of these parses.
other,10-3-J05-1003,ak
attempts to improve upon this initial
<term>
ranking
</term>
, using additional
<term>
features
</term>
#8064A second model then attempts to improve upon this initialranking, using additional features of the tree as evidence.
other,30-12-J05-1003,ak
</term>
which are naturally framed as
<term>
ranking tasks
</term>
, for example ,
<term>
speech recognition
#8331Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed asranking tasks, for example, speech recognition, machine translation, or natural language generation.
other,7-5-J05-1003,ak
We introduce a new method for the
<term>
reranking task
</term>
, based on the
<term>
boosting approach
#8131We introduce a new method for thereranking task, based on the boosting approach to ranking problems described in Freund et al. (1998).
other,16-9-J05-1003,ak
</term>
which takes advantage of the
<term>
sparsity
</term>
of the
<term>
feature space
</term>
in
#8242The article also introduces a new algorithm for the boosting approach which takes advantage of thesparsity of the feature space in the parsing data.
tech,36-12-J05-1003,ak
ranking tasks
</term>
, for example ,
<term>
speech recognition
</term>
,
<term>
machine translation
</term>
#8337Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example,speech recognition, machine translation, or natural language generation.