A <term> language learning experiment </term> showed that <term> assessors </term> can differentiate <term> native from non-native language essays </term> in less than 100 <term> words </term> .
Our <term> algorithm </term> reported more than 99 % <term> accuracy </term> in both <term> language identification </term> and <term> key prediction </term> .
We show that the trained <term> SPR </term> learns to select a <term> sentence plan </term> whose <term> rating </term> on average is only 5 % worse than the <term> top human-ranked sentence plan </term> .
other,18-1-P01-1009,bq This paper presents a <term> formal analysis </term> for a large class of <term> words </term> called <term> alternative markers </term> , which includes <term> other ( than ) </term> , <term> such ( as ) </term> , and <term> besides </term> .
We show that the <term> trainable sentence planner </term> performs better than the <term> rule-based systems </term> and the <term> baselines </term> , and as well as the <term> hand-crafted system </term> .
Surprisingly , learning <term> phrases </term> longer than three <term> words </term> and learning <term> phrases </term> from <term> high-accuracy word-level alignment models </term> does not have a strong impact on performance .
For our purposes , a <term> hub </term> is a <term> node </term> in a <term> graph </term> with <term> in-degree </term> greater than one and <term> out-degree </term> greater than one .
For our purposes , a <term> hub </term> is a <term> node </term> in a <term> graph </term> with <term> in-degree </term> greater than one and <term> out-degree </term> greater than one .
In this paper , we describe a <term> phrase-based unigram model </term> for <term> statistical machine translation </term> that uses a much simpler set of <term> model parameters </term> than similar <term> phrase-based models </term> .
We demonstrate that an approximation of <term> HPSG </term> produces a more effective <term> CFG filter </term> than that of <term> LTAG </term> .
Results indicate that the system yields higher performance than a <term> baseline </term> on all three aspects .
We present controlled experiments showing the <term> WSD </term><term> accuracy </term> of current typical <term> SMT models </term> to be significantly lower than that of all the dedicated <term> WSD models </term> considered .
In this paper we describe a novel <term> data structure </term> for <term> phrase-based statistical machine translation </term> which allows for the <term> retrieval </term> of arbitrarily long <term> phrases </term> while simultaneously using less <term> memory </term> than is required by current <term> decoder </term> implementations .
Using a state-of-the-art <term> Chinese word sense disambiguation model </term> to choose <term> translation candidates </term> for a typical <term> IBM statistical MT system </term> , we find that <term> word sense disambiguation </term> does not yield significantly better <term> translation quality </term> than the <term> statistical machine translation system </term> alone .
We want to illustrate a framework less restrictive than earlier ones by allowing a <term> speaker </term> leeway in forming an <term> utterance </term> about a task and in determining the conversational vehicle to deliver it .
We show that the proposed approach is more describable than other approaches such as those employing a traditional <term> generative phonological approach </term> .
In addition , combination of the <term> training speakers </term> is done by averaging the <term> statistics > </term> of <term> independently trained models </term> rather than the usual pooling of all the <term> speech data </term> from many <term> speakers </term> prior to <term> training </term> .
As each new <term> edge </term> is added to the <term> chart </term> , the algorithm checks only the topmost of the <term> edges </term> adjacent to it , rather than all such <term> edges </term> as in conventional treatments .
A further <term> reduction in the search space </term> is achieved by using <term> semantic </term> rather than <term> syntactic categories </term> on the <term> terminal and non-terminal edges </term> , thereby reducing the amount of <term> ambiguity </term> and thus the number of <term> edges </term> , since only <term> edges </term> with a valid <term> semantic </term> interpretation are ever introduced .
The <term> transfer phase </term> in <term> machine translation ( MT ) systems </term> has been considered to be more complicated than <term> analysis </term> and <term> generation </term> , since it is inherently a conglomeration of individual <term> lexical rules </term> .
hide detail