#8337Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example,speech recognition, machine translation, or natural language generation.
tech,39-12-J05-1003,ak
,
<term>
speech recognition
</term>
,
<term>
machine translation
</term>
, or
<term>
natural language generation
#8340Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example, speech recognition,machine translation, or natural language generation.
other,23-7-J05-1003,ak
evidence from an additional 500,000
<term>
features
</term>
over
<term>
parse trees
</term>
that
#8187The method combined the log-likelihood under a baseline model (that of Collins [1999]) with evidence from an additional 500,000features over parse trees that were not included in the original model.
model,40-4-J05-1003,ak
define a
<term>
derivation
</term>
or a
<term>
generative model
</term>
which takes these
<term>
features
</term>
#8115The strength of our approach is that it allows a tree to be represented as an arbitrary set of features, without concerns about how these features interact or overlap and without the need to define a derivation or agenerative model which takes these features into account.
other,14-3-J05-1003,ak
<term>
ranking
</term>
, using additional
<term>
features
</term>
of the
<term>
tree
</term>
as evidence
#8068A second model then attempts to improve upon this initial ranking, using additionalfeatures of the tree as evidence.
other,30-12-J05-1003,ak
</term>
which are naturally framed as
<term>
ranking tasks
</term>
, for example ,
<term>
speech recognition
#8331Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed asranking tasks, for example, speech recognition, machine translation, or natural language generation.
other,16-2-J05-1003,ak
input sentence
</term>
, with associated
<term>
probabilities
</term>
that define an initial
<term>
ranking
#8044The base parser produces a set of candidate parses for each input sentence, with associatedprobabilities that define an initial ranking of these parses.
other,11-2-J05-1003,ak
<term>
candidate parses
</term>
for each
<term>
input sentence
</term>
, with associated
<term>
probabilities
#8039The base parser produces a set of candidate parses for eachinput sentence, with associated probabilities that define an initial ranking of these parses.
tech,11-1-J05-1003,ak
which rerank the output of an existing
<term>
probabilistic parser
</term>
. The
<term>
base parser
</term>
produces
#8025This article considers approaches which rerank the output of an existingprobabilistic parser.
measure(ment),14-8-J05-1003,ak
</term>
, a 13 % relative decrease in
<term>
F-measure error
</term>
over the
<term>
baseline model ’s
</term>
#8214The new model achieved 89.75% F-measure, a 13% relative decrease inF-measure error over the baseline model’s score of 88.2%.
other,7-2-J05-1003,ak
base parser
</term>
produces a set of
<term>
candidate parses
</term>
for each
<term>
input sentence
</term>
#8035The base parser produces a set ofcandidate parses for each input sentence, with associated probabilities that define an initial ranking of these parses.
other,19-4-J05-1003,ak
represented as an arbitrary set of
<term>
features
</term>
, without concerns about how these
#8094The strength of our approach is that it allows a tree to be represented as an arbitrary set offeatures, without concerns about how these features interact or overlap and without the need to define a derivation or a generative model which takes these features into account.
tech,8-12-J05-1003,ak
experiments in this article are on
<term>
natural language parsing ( NLP )
</term>
, the approach should be applicable
#8309Although the experiments in this article are onnatural language parsing ( NLP ), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example, speech recognition, machine translation, or natural language generation.
tech,21-11-J05-1003,ak
simplicity and efficiency — to work on
<term>
feature selection methods
</term>
within
<term>
log-linear ( maximum-entropy
#8291We argue that the method is an appealing alternative—in terms of both simplicity and efficiency—to work onfeature selection methods within log-linear (maximum-entropy) models.
tech,43-12-J05-1003,ak
<term>
machine translation
</term>
, or
<term>
natural language generation
</term>
. We present a novel method for discovering
#8344Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example, speech recognition, machine translation, ornatural language generation.
tech,23-12-J05-1003,ak
should be applicable to many other
<term>
NLP problems
</term>
which are naturally framed as
<term>
#8324Although the experiments in this article are on natural language parsing (NLP), the approach should be applicable to many otherNLP problems which are naturally framed as ranking tasks, for example, speech recognition, machine translation, or natural language generation.
other,25-7-J05-1003,ak
additional 500,000
<term>
features
</term>
over
<term>
parse trees
</term>
that were not included in the original
#8189The method combined the log-likelihood under a baseline model (that of Collins [1999]) with evidence from an additional 500,000 features overparse trees that were not included in the original model.
tech,1-2-J05-1003,ak
<term>
probabilistic parser
</term>
. The
<term>
base parser
</term>
produces a set of
<term>
candidate
#8029Thebase parser produces a set of candidate parses for each input sentence, with associated probabilities that define an initial ranking of these parses.
tech,3-6-J05-1003,ak
Freund et al. ( 1998 ) . We apply the
<term>
boosting method
</term>
to parsing the
<term>
Wall Street Journal
#8154We apply theboosting method to parsing the Wall Street Journal treebank.
other,4-7-J05-1003,ak
treebank
</term>
. The method combined the
<term>
log-likelihood under a baseline model
</term>
( that of Collins [ 1999 ] ) with
#8168The method combined thelog-likelihood under a baseline model (that of Collins [1999]) with evidence from an additional 500,000 features over parse trees that were not included in the original model.