<term> Communicator </term> participants are using . In this presentation , we describe the
elicited from duplicating the experiment using <term> machine translation output </term> .
<term> word string </term> has been obtained by using a different <term> LM </term> . Actually ,
surprisingly close to what can be achieved using conventional <term> word-trigram recognition
performance gains from the <term> data </term> by using <term> class-dependent interpolation </term>
NIST score </term> demonstrated the effect of using an out-of-domain <term> bilingual corpus </term>
<term> Named Entity ( NE ) tagging </term> using <term> concept-based seeds </term> and <term>
<term> source interval projections </term> using an underlying <term> word alignment </term>
subcategorization frame ( SCF ) </term> distributions using the <term> Information Bottleneck </term> and
derived by <term> decision tree learning </term> using real <term> dialogue data </term> collected
proprietary <term> Arabic stemmer </term> built using <term> rules </term> , <term> affix lists </term>
</term> based on the results . The evaluation using another 23 subjects showed that the proposed
of a <term> term aggregation system </term> using each author 's text as a coherent <term>
</term> in <term> unannotated text </term> by using a fully automatic sequence of <term> preprocessing
patterns </term> in <term> translation data </term> using <term> part-of-speech tag sequences </term>
sense disambiguation performance </term> , using standard <term> WSD evaluation methodology
improve upon this initial <term> ranking </term> , using additional <term> features </term> of the <term>
<term> phrases </term> while simultaneously using less <term> memory </term> than is required
efficient <term> decoder </term> and show that using these <term> tree-based models </term> in combination
outputs </term> of our <term> MT system </term> using the <term> NIST and Bleu automatic MT evaluation
hide detail