probabilities
</term>
that define an initial
<term>
ranking
</term>
of these
<term>
parses
</term>
. A second
#8049The base parser produces a set of candidate parses for each input sentence, with associated probabilities that define an initialranking of these parses.
tech,6-9-J05-1003,ak
The article also introduces a new
<term>
algorithm
</term>
for the
<term>
boosting approach
</term>
#8232The article also introduces a newalgorithm for the boosting approach which takes advantage of the sparsity of the feature space in the parsing data.
other,19-4-J05-1003,ak
represented as an arbitrary set of
<term>
features
</term>
, without concerns about how these
#8094The strength of our approach is that it allows a tree to be represented as an arbitrary set offeatures, without concerns about how these features interact or overlap and without the need to define a derivation or a generative model which takes these features into account.
other,45-4-J05-1003,ak
generative model
</term>
which takes these
<term>
features
</term>
into account . We introduce a new
#8120The strength of our approach is that it allows a tree to be represented as an arbitrary set of features, without concerns about how these features interact or overlap and without the need to define a derivation or a generative model which takes thesefeatures into account.
model,2-8-J05-1003,ak
original
<term>
model
</term>
. The new
<term>
model
</term>
achieved 89.75 %
<term>
F-measure
</term>
#8202The newmodel achieved 89.75% F-measure, a 13% relative decrease in F-measure error over the baseline model’s score of 88.2%.
tech,21-11-J05-1003,ak
simplicity and efficiency — to work on
<term>
feature selection methods
</term>
within
<term>
log-linear ( maximum-entropy
#8291We argue that the method is an appealing alternative—in terms of both simplicity and efficiency—to work onfeature selection methods within log-linear (maximum-entropy) models.
tech,13-5-J05-1003,ak
reranking task
</term>
, based on the
<term>
boosting approach to ranking problems
</term>
described in Freund et al. ( 1998
#8137We introduce a new method for the reranking task, based on theboosting approach to ranking problems described in Freund et al. (1998).
model,25-11-J05-1003,ak
feature selection methods
</term>
within
<term>
log-linear ( maximum-entropy ) models
</term>
. Although the experiments in this
#8295We argue that the method is an appealing alternative—in terms of both simplicity and efficiency—to work on feature selection methods withinlog-linear ( maximum-entropy ) models.
measure(ment),14-8-J05-1003,ak
</term>
, a 13 % relative decrease in
<term>
F-measure error
</term>
over the
<term>
baseline model ’s
</term>
#8214The new model achieved 89.75% F-measure, a 13% relative decrease inF-measure error over the baseline model’s score of 88.2%.
other,19-9-J05-1003,ak
of the
<term>
sparsity
</term>
of the
<term>
feature space
</term>
in the
<term>
parsing data
</term>
.
#8245The article also introduces a new algorithm for the boosting approach which takes advantage of the sparsity of thefeature space in the parsing data.
tech,8-10-J05-1003,ak
significant efficiency gains for the new
<term>
algorithm
</term>
over the obvious implementation of
#8260Experiments show significant efficiency gains for the newalgorithm over the obvious implementation of the boosting approach.
other,30-12-J05-1003,ak
</term>
which are naturally framed as
<term>
ranking tasks
</term>
, for example ,
<term>
speech recognition
#8331Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed asranking tasks, for example, speech recognition, machine translation, or natural language generation.
lr,8-6-J05-1003,ak
boosting method
</term>
to parsing the
<term>
Wall Street Journal treebank
</term>
. The method combined the
<term>
log-likelihood
#8159We apply the boosting method to parsing theWall Street Journal treebank.
other,11-2-J05-1003,ak
<term>
candidate parses
</term>
for each
<term>
input sentence
</term>
, with associated
<term>
probabilities
#8039The base parser produces a set of candidate parses for eachinput sentence, with associated probabilities that define an initial ranking of these parses.
tech,11-1-J05-1003,ak
which rerank the output of an existing
<term>
probabilistic parser
</term>
. The
<term>
base parser
</term>
produces
#8025This article considers approaches which rerank the output of an existingprobabilistic parser.
tech,39-12-J05-1003,ak
,
<term>
speech recognition
</term>
,
<term>
machine translation
</term>
, or
<term>
natural language generation
#8340Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example, speech recognition,machine translation, or natural language generation.
measure(ment),6-8-J05-1003,ak
<term>
model
</term>
achieved 89.75 %
<term>
F-measure
</term>
, a 13 % relative decrease in
<term>
#8206The new model achieved 89.75%F-measure, a 13% relative decrease in F-measure error over the baseline model’s score of 88.2%.
tech,3-6-J05-1003,ak
Freund et al. ( 1998 ) . We apply the
<term>
boosting method
</term>
to parsing the
<term>
Wall Street Journal
#8154We apply theboosting method to parsing the Wall Street Journal treebank.
other,23-9-J05-1003,ak
the
<term>
feature space
</term>
in the
<term>
parsing data
</term>
. Experiments show significant efficiency
#8249The article also introduces a new algorithm for the boosting approach which takes advantage of the sparsity of the feature space in theparsing data.
tech,36-12-J05-1003,ak
ranking tasks
</term>
, for example ,
<term>
speech recognition
</term>
,
<term>
machine translation
</term>
#8337Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example,speech recognition, machine translation, or natural language generation.