|
mimics the behavior of the
<term>
oracle
</term>
|
using
|
a
<term>
neural network
</term>
or a
<term>
decision
|
#1163
We suggest a method that mimics the behavior of the oracleusing a neural network or a decision tree. |
|
in the search space
</term>
is achieved by
|
using
|
<term>
semantic
</term>
rather than
<term>
syntactic
|
#17713
A further reduction in the search space is achieved by using semantic rather than syntactic categories on the terminal and non-terminal edges, thereby reducing the amount of ambiguity and thus the number of edges, since only edges with a valid semantic interpretation are ever introduced. |
|
subcategorization frame ( SCF )
</term>
distributions
|
using
|
the
<term>
Information Bottleneck
</term>
and
|
#3916
We describe a new approach which involves clustering subcategorization frame (SCF) distributions using the Information Bottleneck and nearest neighbour methods. |
|
</term>
based on the results . The evaluation
|
using
|
another 23 subjects showed that the proposed
|
#5713
The evaluation using another 23 subjects showed that the proposed method could effectively generate proper referring expressions. |
|
</term>
. We show that this task can be done
|
using
|
<term>
bilingual parallel corpora
</term>
,
|
#9675
We show that this task can be done using bilingual parallel corpora, a much more commonly available resource. |
|
shown that these results can be improved
|
using
|
a bigger and a more homogeneous
<term>
corpus
|
#11293
Finally, we have shown that these results can be improved using a bigger and a more homogeneous corpus to train, that is, a bigger corpus written by one unique author. |
|
</term>
in
<term>
unannotated text
</term>
by
|
using
|
a fully automatic sequence of
<term>
preprocessing
|
#7083
Furthermore, we present a standalone system that resolves pronouns in unannotated text by using a fully automatic sequence of preprocessing modules that mimics the manual annotation process. |
|
</term>
and
<term>
linguistic pattern
</term>
. By
|
using
|
them , we can automatically extract such
|
#11442
By using them, we can automatically extract such sentences that express opinion. |
|
bilingual parallel corpus
</term>
to be ranked
|
using
|
<term>
translation probabilities
</term>
,
|
#9733
We define a paraphrase probability that allows paraphrases extracted from a bilingual parallel corpus to be ranked using translation probabilities, and show how it can be refined to take contextual information into account. |
|
for
<term>
speaker adaptation ( SA )
</term>
|
using
|
the new
<term>
SI corpus
</term>
and a small
|
#17132
Second, we show a significant improvement for speaker adaptation (SA)using the new SI corpus and a small amount of speech from the new (target) speaker. |
|
automatically from
<term>
raw text
</term>
. Experiments
|
using
|
the
<term>
SemCor
</term>
and
<term>
Senseval-3
|
#11025
Experiments using the SemCor and Senseval-3 data sets demonstrate that our ensembles yield significantly better results when compared with state-of-the-art. |
|
the sum of each
<term>
character
</term>
. By
|
using
|
commands or
<term>
rules
</term>
which are
|
#12312
By using commands or rules which are defined to facilitate the construction of format expected or some mathematical expressions, elaborate and pretty documents can be successfully obtained. |
|
sense disambiguation performance
</term>
,
|
using
|
standard
<term>
WSD evaluation methodology
|
#7814
We present the first known empirical test of an increasingly common speculative claim, by evaluating a representative Chinese-to-English SMT model directly on word sense disambiguation performance, using standard WSD evaluation methodology and datasets from the Senseval-3 Chinese lexical sample task. |
|
<term>
word string
</term>
has been obtained by
|
using
|
a different
<term>
LM
</term>
. Actually ,
|
#1110
The oracle knows the reference word string and selects the word string with the best performance (typically, word or semantic error rate) from a list of word strings, where each word string has been obtained by using a different LM. |
|
describe the methods and hardware that we are
|
using
|
to produce a real-time demonstration of
|
#16874
We describe the methods and hardware that we are using to produce a real-time demonstration of an integrated Spoken Language System. |
|
corpora
</term>
is presented which involves
|
using
|
a
<term>
statistical POS tagger
</term>
in
|
#19958
A novel method for adding linguistic annotation to corpora is presented which involves using a statistical POS tagger in conjunction with unsupervised structure finding methods to derive notions of noun group, verb group, and so on which is inherently extensible to more sophisticated annotation, and does not require a pre-tagged corpus to fit. |
|
</term>
, this paper proposes new methods
|
using
|
<term>
m-th order Markov chain model
</term>
|
#20680
In order to judge three types of the errors, which are characters wrongly substituted, deleted or inserted in a Japanese bunsetsu and an English word, and to correct these errors, this paper proposes new methods using m-th order Markov chain model for Japanese kanji-kana characters and English alphabets, assuming that Markov probability of a correct chain of syllables or kanji-kana characters is greater than that of erroneous chains. |
|
patterns
</term>
in
<term>
translation data
</term>
|
using
|
<term>
part-of-speech tag sequences
</term>
|
#7639
We describe a method for identifying systematic patterns in translation datausing part-of-speech tag sequences. |
|
a
<term>
token classification task
</term>
,
|
using
|
various
<term>
tagging strategies
</term>
to
|
#10813
There are several approaches that model information extraction as a token classification task, using various tagging strategies to combine multiple tokens. |
|
(
<term>
anaphora
</term>
) . This method of
|
using
|
<term>
expectations
</term>
to aid the understanding
|
#13102
This method of using expectations to aid the understanding of scruffy texts has been incorporated into a working computer program called NOMAD, which understands scruffy texts in the domain of Navy messages. |