To support engaging human users in robust , <term> mixed-initiative speech dialogue interactions </term> which reach beyond current capabilities in <term> dialogue systems </term> , the <term> DARPA Communicator program </term> [ 1 ] is funding the development of a <term> distributed message-passing infrastructure </term> for <term> dialogue systems </term> which all <term> Communicator </term> participants are using .
We tested this to see if similar criteria could be elicited from duplicating the experiment using <term> machine translation output </term> .
The <term> oracle </term> knows the <term> reference word string </term> and selects the <term> word string </term> with the best <term> performance </term> ( typically , <term> word or semantic error rate </term> ) from a list of <term> word strings </term> , where each <term> word string </term> has been obtained by using a different <term> LM </term> .
Actually , the <term> oracle </term> acts like a <term> dynamic combiner </term> with <term> hard decisions </term> using the <term> reference </term> .
We suggest a method that mimics the behavior of the <term> oracle </term> using a <term> neural network </term> or a <term> decision tree </term> .
The method combines <term> domain independent acoustic models </term> with off-the-shelf <term> classifiers </term> to give <term> utterance classification performance </term> that is surprisingly close to what can be achieved using conventional <term> word-trigram recognition </term> requiring <term> manual transcription </term> .
In this paper , we show how <term> training data </term> can be supplemented with <term> text </term> from the <term> web </term> filtered to match the <term> style </term> and/or <term> topic </term> of the target <term> recognition task </term> , but also that it is possible to get bigger performance gains from the <term> data </term> by using <term> class-dependent interpolation </term> of <term> N-grams </term> .
The two <term> evaluation measures </term> of the <term> BLEU score </term> and the <term> NIST score </term> demonstrated the effect of using an out-of-domain <term> bilingual corpus </term> and the possibility of using the <term> language model </term> .
The two <term> evaluation measures </term> of the <term> BLEU score </term> and the <term> NIST score </term> demonstrated the effect of using an out-of-domain <term> bilingual corpus </term> and the possibility of using the <term> language model </term> .
A novel <term> bootstrapping approach </term> to <term> Named Entity ( NE ) tagging </term> using <term> concept-based seeds </term> and <term> successive learners </term> is presented .
During <term> training </term> , the <term> blocks </term> are learned from <term> source interval projections </term> using an underlying <term> word alignment </term> .
We describe a new approach which involves clustering <term> subcategorization frame ( SCF ) </term> distributions using the <term> Information Bottleneck </term> and <term> nearest neighbour </term> methods .
Moreover , the <term> models </term> are automatically derived by <term> decision tree learning </term> using real <term> dialogue data </term> collected by the <term> system </term> .
Our <term> resource-frugal approach </term> results in 87.5 % <term> agreement </term> with a state of the art , proprietary <term> Arabic stemmer </term> built using <term> rules </term> , <term> affix lists </term> , and <term> human annotated text </term> , in addition to an <term> unsupervised component </term> .
<term> Task-based evaluation </term> using <term> Arabic information retrieval </term> indicates an improvement of 22-38 % in <term> average precision </term> over <term> unstemmed text </term> , and 96 % of the performance of the proprietary <term> stemmer </term> above .
The evaluation using another 23 subjects showed that the proposed method could effectively generate proper <term> referring expressions </term> .
This paper proposes a new methodology to improve the <term> accuracy </term> of a <term> term aggregation system </term> using each author 's text as a coherent <term> corpus </term> .
Furthermore , we present a standalone system that resolves <term> pronouns </term> in <term> unannotated text </term> by using a fully automatic sequence of <term> preprocessing modules </term> that mimics the manual <term> annotation process </term> .
We describe a <term> method </term> for identifying systematic <term> patterns </term> in <term> translation data </term> using <term> part-of-speech tag sequences </term> .
We present the first known <term> empirical test </term> of an increasingly common speculative claim , by evaluating a representative <term> Chinese-to-English SMT model </term> directly on <term> word sense disambiguation performance </term> , using standard <term> WSD evaluation methodology </term> and <term> datasets </term> from the <term> Senseval-3 Chinese lexical sample task </term> .
hide detail