other,10-4-J05-1003,bq |
approach
</term>
is that it allows a
<term>
|
tree
|
</term>
to be represented as an arbitrary
|
#8720
The strength of our approach is that it allows atree to be represented as an arbitrary set of features, without concerns about how these features interact or overlap and without the need to define a derivation or a generative model which takes these features into account. |
tech,6-6-J05-1003,bq |
the
<term>
boosting method
</term>
to
<term>
|
parsing
|
</term>
the
<term>
Wall Street Journal treebank
|
#8792
We apply the boosting method toparsing the Wall Street Journal treebank. |
tech,2-2-J05-1003,bq |
probabilistic parser
</term>
. The base
<term>
|
parser
|
</term>
produces a set of
<term>
candidate
|
#8665
The baseparser produces a set of candidate parses for each input sentence, with associated probabilities that define an initial ranking of these parses. |
other,21-2-J05-1003,bq |
probabilities
</term>
that define an initial
<term>
|
ranking
|
</term>
of these
<term>
parses
</term>
. A second
|
#8684
The base parser produces a set of candidate parses for each input sentence, with associated probabilities that define an initialranking of these parses. |
other,19-4-J05-1003,bq |
represented as an arbitrary set of
<term>
|
features
|
</term>
, without concerns about how these
|
#8729
The strength of our approach is that it allows a tree to be represented as an arbitrary set offeatures, without concerns about how these features interact or overlap and without the need to define a derivation or a generative model which takes these features into account. |
tech,1-7-J05-1003,bq |
Street Journal treebank
</term>
. The
<term>
|
method
|
</term>
combined the
<term>
log-likelihood
</term>
|
#8800
Themethod combined the log-likelihood under a baseline model (that of Collins [1999]) with evidence from an additional 500,000 features over parse trees that were not included in the original model. |
other,12-2-J05-1003,bq |
candidate parses
</term>
for each input
<term>
|
sentence
|
</term>
, with associated
<term>
probabilities
|
#8675
The base parser produces a set of candidate parses for each inputsentence, with associated probabilities that define an initial ranking of these parses. |
tech,3-6-J05-1003,bq |
al. ( 1998 )
</term>
. We apply the
<term>
|
boosting method
|
</term>
to
<term>
parsing
</term>
the
<term>
Wall
|
#8789
We apply theboosting method to parsing the Wall Street Journal treebank. |
tech,8-12-J05-1003,bq |
experiments in this article are on
<term>
|
natural language parsing ( NLP )
|
</term>
, the
<term>
approach
</term>
should
|
#8944
Although the experiments in this article are onnatural language parsing ( NLP ), the approach should be applicable to many other NLP problems which are naturally framed as ranking tasks, for example, speech recognition, machine translation, or natural language generation. |
other,16-5-J05-1003,bq |
the
<term>
boosting approach
</term>
to
<term>
|
ranking problems
|
</term>
described in
<term>
Freund et al. (
|
#8775
We introduce a new method for the reranking task, based on the boosting approach toranking problems described in Freund et al. (1998). |
tech,2-3-J05-1003,bq |
these
<term>
parses
</term>
. A second
<term>
|
model
|
</term>
then attempts to improve upon this
|
#8691
A secondmodel then attempts to improve upon this initial ranking, using additional features of the tree as evidence. |
tech,7-5-J05-1003,bq |
introduce a new
<term>
method
</term>
for the
<term>
|
reranking task
|
</term>
, based on the
<term>
boosting approach
|
#8766
We introduce a new method for thereranking task, based on the boosting approach to ranking problems described in Freund et al. (1998). |
tech,6-9-J05-1003,bq |
The article also introduces a new
<term>
|
algorithm
|
</term>
for the
<term>
boosting approach
</term>
|
#8867
The article also introduces a newalgorithm for the boosting approach which takes advantage of the sparsity of the feature space in the parsing data. |
measure(ment),6-8-J05-1003,bq |
<term>
model
</term>
achieved 89.75 %
<term>
|
F-measure
|
</term>
, a 13 % relative decrease in
<term>
|
#8841
The new model achieved 89.75%F-measure, a 13% relative decrease in F-measure error over the baseline model’s score of 88.2%. |
other,26-4-J05-1003,bq |
, without concerns about how these
<term>
|
features
|
</term>
interact or overlap and without the
|
#8736
The strength of our approach is that it allows a tree to be represented as an arbitrary set of features, without concerns about how thesefeatures interact or overlap and without the need to define a derivation or a generative model which takes these features into account. |
other,45-4-J05-1003,bq |
generative model
</term>
which takes these
<term>
|
features
|
</term>
into account . We introduce a new
|
#8755
The strength of our approach is that it allows a tree to be represented as an arbitrary set of features, without concerns about how these features interact or overlap and without the need to define a derivation or a generative model which takes thesefeatures into account. |
other,12-7-J05-1003,bq |
<term>
baseline model
</term>
( that of
<term>
|
Collins [ 1999 ]
|
</term>
) with evidence from an additional
|
#8811
The method combined the log-likelihood under a baseline model (that ofCollins [ 1999 ]) with evidence from an additional 500,000 features over parse trees that were not included in the original model. |
tech,4-5-J05-1003,bq |
</term>
into account . We introduce a new
<term>
|
method
|
</term>
for the
<term>
reranking task
</term>
|
#8763
We introduce a newmethod for the reranking task, based on the boosting approach to ranking problems described in Freund et al. (1998). |
other,25-7-J05-1003,bq |
additional 500,000
<term>
features
</term>
over
<term>
|
parse trees
|
</term>
that were not included in the original
|
#8824
The method combined the log-likelihood under a baseline model (that of Collins [1999]) with evidence from an additional 500,000 features overparse trees that were not included in the original model. |
other,23-9-J05-1003,bq |
of the feature space
</term>
in the
<term>
|
parsing data
|
</term>
. Experiments show significant efficiency
|
#8884
The article also introduces a new algorithm for the boosting approach which takes advantage of the sparsity of the feature space in theparsing data. |