tech,13-2-P03-1050,ak machine translation </term> and it uses an <term> English stemmer </term> and a <term> small ( 10K sentences
other,22-4-P03-1050,ak to a desired <term> domain </term> or <term> genre </term> . Examples and results will be given
other,7-5-P03-1050,ak Examples and results will be given for <term> Arabic </term> , but the approach is applicable
measure(ment),7-6-P03-1050,ak resource-frugal approach </term> results in 87.5 % <term> agreement </term> with a state of the art , proprietary
other,16-7-P03-1050,ak <term> average precision </term> over <term> unstemmed text </term> , and 96 % of the performance of
tech,16-6-P03-1050,ak with a state of the art , proprietary <term> Arabic stemmer </term> built using <term> rules </term> , <term>
model,20-6-P03-1050,ak <term> Arabic stemmer </term> built using <term> rules </term> , <term> affix lists </term> , and <term>
tech,10-1-P03-1050,ak learning approach </term> to building a <term> non-English ( Arabic ) stemmer </term> . The <term> stemming model </term> is
lr,27-2-P03-1050,ak parallel corpus </term> as its sole <term> training resources </term> . No <term> parallel text </term> is
other,20-4-P03-1050,ak allowing it to adapt to a desired <term> domain </term> or <term> genre </term> . Examples and
tech,6-2-P03-1050,ak <term> stemming model </term> is based on <term> statistical machine translation </term> and it uses an <term> English stemmer
tech,3-7-P03-1050,ak <term> Task-based evaluation </term> using <term> Arabic information retrieval </term> indicates an improvement of 22-38
lr,26-6-P03-1050,ak </term> , <term> affix lists </term> , and <term> human annotated text </term> , in addition to an <term> unsupervised
model,1-2-P03-1050,ak non-English ( Arabic ) stemmer </term> . The <term> stemming model </term> is based on <term> statistical machine
tech,4-1-P03-1050,ak users </term> . This paper presents an <term> unsupervised learning approach </term> to building a <term> non-English (
measure(ment),13-7-P03-1050,ak indicates an improvement of 22-38 % in <term> average precision </term> over <term> unstemmed text </term> ,
tech,28-7-P03-1050,ak the performance of the proprietary <term> stemmer </term> above . We approximate <term> Arabic
lr,22-6-P03-1050,ak </term> built using <term> rules </term> , <term> affix lists </term> , and <term> human annotated text </term>
tech,19-5-P03-1050,ak any <term> language </term> that needs <term> affix removal </term> . Our <term> resource-frugal approach
tech,11-4-P03-1050,ak can be used to further improve the <term> stemmer </term> by allowing it to adapt to a desired
hide detail