tech,2-3-J05-4003,bq Using this <term> approach </term> , we extract <term> parallel data </term> from large <term> Chinese , Arabic , and English non-parallel newspaper corpora </term> .
lr,10-3-J05-4003,bq Using this <term> approach </term> , we extract <term> parallel data </term> from large <term> Chinese , Arabic , and English non-parallel newspaper corpora </term> .
lr,10-1-J05-4003,bq We present a novel <term> method </term> for <term> discovering parallel sentences </term> in <term> comparable , non-parallel corpora </term> .
tech,6-1-J05-4003,bq We present a novel <term> method </term> for <term> discovering parallel sentences </term> in <term> comparable , non-parallel corpora </term> .
other,11-6-J05-4003,bq Thus , our method can be applied with great benefit to <term> language pairs </term> for which only scarce <term> resources </term> are available .
tech,3-2-J05-4003,bq We train a <term> maximum entropy classifier </term> that , given a pair of <term> sentences </term> , can reliably determine whether or not they are <term> translations </term> of each other .
tech,4-1-J05-4003,bq We present a novel <term> method </term> for <term> discovering parallel sentences </term> in <term> comparable , non-parallel corpora </term> .
tech,6-5-J05-4003,bq We also show that a good-quality <term> MT system </term> can be built from scratch by starting with a very small <term> parallel corpus </term> ( 100,000 <term> words </term> ) and exploiting a large <term> non-parallel corpus </term> .
lr,29-5-J05-4003,bq We also show that a good-quality <term> MT system </term> can be built from scratch by starting with a very small <term> parallel corpus </term> ( 100,000 <term> words </term> ) and exploiting a large <term> non-parallel corpus </term> .
lr,19-5-J05-4003,bq We also show that a good-quality <term> MT system </term> can be built from scratch by starting with a very small <term> parallel corpus </term> ( 100,000 <term> words </term> ) and exploiting a large <term> non-parallel corpus </term> .
lr,6-3-J05-4003,bq Using this <term> approach </term> , we extract <term> parallel data </term> from large <term> Chinese , Arabic , and English non-parallel newspaper corpora </term> .
measure(ment),3-4-J05-4003,bq We evaluate the <term> quality of the extracted data </term> by showing that it improves the performance of a state-of-the-art <term> statistical machine translation system </term> .
lr,17-6-J05-4003,bq Thus , our method can be applied with great benefit to <term> language pairs </term> for which only scarce <term> resources </term> are available .
other,12-2-J05-4003,bq We train a <term> maximum entropy classifier </term> that , given a pair of <term> sentences </term> , can reliably determine whether or not they are <term> translations </term> of each other .
tech,18-4-J05-4003,bq We evaluate the <term> quality of the extracted data </term> by showing that it improves the performance of a state-of-the-art <term> statistical machine translation system </term> .
other,22-2-J05-4003,bq We train a <term> maximum entropy classifier </term> that , given a pair of <term> sentences </term> , can reliably determine whether or not they are <term> translations </term> of each other .
other,23-5-J05-4003,bq We also show that a good-quality <term> MT system </term> can be built from scratch by starting with a very small <term> parallel corpus </term> ( 100,000 <term> words </term> ) and exploiting a large <term> non-parallel corpus </term> .
hide detail