other,27-3-H01-1041,bq case markers </term> , relatively <term> free word order </term> , and frequent omissions of
tech,7-4-H01-1041,bq quality <term> translation </term> via <term> word sense disambiguation </term> and accurate
tech,12-4-H01-1041,bq disambiguation </term> and accurate <term> word order generation </term> of the <term> target
other,8-10-H01-1042,bq Additionally , they were asked to mark the <term> word </term> at which they made this decision
other,4-3-H01-1058,bq <term> oracle </term> knows the <term> reference word string </term> and selects the <term> word
other,10-3-H01-1058,bq word string </term> and selects the <term> word string </term> with the best <term> performance
measure(ment),19-3-H01-1058,bq <term> performance </term> ( typically , <term> word or semantic error rate </term> ) from a list
other,29-3-H01-1058,bq error rate </term> ) from a list of <term> word strings </term> , where each <term> word string
other,34-3-H01-1058,bq <term> word strings </term> , where each <term> word string </term> has been obtained by using
model,24-3-P01-1004,bq </term> superior to any of the tested <term> word N-gram models </term> . Further , in their
other,3-3-P01-1008,bq approach yields <term> phrasal and single word lexical paraphrases </term> as well as <term>
measure(ment),20-3-N03-1018,bq significantly reduce <term> character and word error rate </term> , and provide evaluation
other,66-1-N03-1033,bq ) fine-grained modeling of <term> unknown word features </term> . Using these ideas together
tech,6-1-N03-2017,bq <term> syntax-based constraint </term> for <term> word alignment </term> , known as the <term> cohesion
model,14-4-N03-2036,bq projections </term> using an underlying <term> word alignment </term> . We show experimental
other,11-1-P03-1051,bq </term> by a <term> model </term> that a <term> word </term> consists of a sequence of <term> morphemes
tech,22-2-P03-1051,bq algorithm </term> to build the <term> Arabic word segmenter </term> from a large <term> unsegmented
other,20-5-P03-1051,bq <term> stems </term> from a 155 million <term> word </term><term> unsegmented corpus </term> ,
tech,2-6-P03-1051,bq corpus </term> . The resulting <term> Arabic word segmentation system </term> achieves around
other,19-6-P03-1051,bq test corpus </term> containing 28,449 <term> word tokens </term> . We believe this is a state-of-the-art
hide detail