The paper also proposes <term> rule-reduction algorithm </term> applying <term> mutual information </term> to reduce the <term> error-correction rules </term> . Our <term> algorithm </term> reported more than 99 % <term> accuracy </term> in both <term> language identification </term> and <term> key prediction </term> .
We present an <term> unsupervised learning algorithm </term> for <term> identification of paraphrases </term> from a <term> corpus of multiple English translations </term> of the same <term> source text </term> . Our approach yields <term> phrasal and single word lexical paraphrases </term> as well as <term> syntactic paraphrases </term> .
We provide a <term> logical definition </term> of <term> Minimalist grammars </term> , that are <term> Stabler 's formalization </term> of <term> Chomsky 's minimalist program </term> . Our <term> logical definition </term> leads to a neat relation to <term> categorial grammar </term> , ( yielding a treatment of <term> Montague semantics </term> ) , a <term> parsing-as-deduction </term> in a <term> resource sensitive logic </term> , and a <term> learning algorithm </term> from <term> structured data </term> ( based on a <term> typing-algorithm </term> and <term> type-unification </term> ) .
Within our framework , we carry out a large number of experiments to understand better and explain why <term> phrase-based models </term> outperform <term> word-based models </term> . Our empirical results , which hold for all examined <term> language pairs </term> , suggest that the highest levels of performance can be obtained through relatively simple means : <term> heuristic learning </term> of <term> phrase translations </term> from <term> word-based alignments </term> and <term> lexical weighting </term> of <term> phrase translations </term> .
We present an application of <term> ambiguity packing and stochastic disambiguation techniques </term> for <term> Lexical-Functional Grammars ( LFG ) </term> to the domain of <term> sentence condensation </term> . Our <term> system </term> incorporates a <term> linguistic parser/generator </term> for <term> LFG </term> , a <term> transfer component </term> for <term> parse reduction </term> operating on <term> packed parse forests </term> , and a <term> maximum-entropy model </term> for <term> stochastic output selection </term> .
We apply a <term> decision tree based approach </term> to <term> pronoun resolution </term> in <term> spoken dialogue </term> . Our <term> system </term> deals with <term> pronouns </term> with <term> NP - and non-NP-antecedents </term> .
Examples and results will be given for <term> Arabic </term> , but the approach is applicable to any <term> language </term> that needs <term> affix removal </term> . Our <term> resource-frugal approach </term> results in 87.5 % <term> agreement </term> with a state of the art , proprietary <term> Arabic stemmer </term> built using <term> rules </term> , <term> affix lists </term> , and <term> human annotated text </term> , in addition to an <term> unsupervised component </term> .
We approximate <term> Arabic 's rich morphology </term> by a <term> model </term> that a <term> word </term> consists of a sequence of <term> morphemes </term> in the <term> pattern </term><term> prefix * - stem-suffix * </term> ( * denotes zero or more occurrences of a <term> morpheme </term> ) . Our method is seeded by a small <term> manually segmented Arabic corpus </term> and uses it to bootstrap an <term> unsupervised algorithm </term> to build the <term> Arabic word segmenter </term> from a large <term> unsegmented Arabic corpus </term> .
In this paper , we evaluate an approach to automatically acquire <term> sense-tagged training data </term> from <term> English-Chinese parallel corpora </term> , which are then used for disambiguating the <term> nouns </term> in the <term> SENSEVAL-2 English lexical sample task </term> . Our investigation reveals that this <term> method of acquiring sense-tagged data </term> is promising .
On a subset of the most difficult <term> SENSEVAL-2 nouns </term> , the <term> accuracy </term> difference between the two approaches is only 14.0 % , and the difference could narrow further to 6.5 % if we disregard the advantage that <term> manually sense-tagged data </term> have in their <term> sense coverage </term> . Our analysis also highlights the importance of the issue of <term> domain dependence </term> in evaluating <term> WSD programs </term> .
We tested the <term> clustering and filtering processes </term> on <term> electronic newsgroup discussions </term> , and evaluated their performance by means of two experiments : coarse-level <term> clustering </term> and simple <term> information retrieval </term> . Our evaluation shows that our <term> filtering mechanism </term> has a significant positive effect on both tasks .
We evaluate the proposed methods through several <term> transliteration/back transliteration </term> experiments for <term> English/Chinese and English/Japanese language pairs </term> . Our study reveals that the proposed method not only reduces an extensive system development effort but also improves the <term> transliteration accuracy </term> significantly .
This paper proposes a new methodology to improve the <term> accuracy </term> of a <term> term aggregation system </term> using each author 's text as a coherent <term> corpus </term> . Our approach is based on the idea that one person tends to use one <term> expression </term> for one <term> meaning </term> .
According to our assumption , most of the words with similar <term> context features </term> in each author 's <term> corpus </term> tend not to be <term> synonymous expressions </term> . Our proposed method improves the <term> accuracy </term> of our <term> term aggregation system </term> , showing that our approach is successful .
We report the performance of the <term> MBR decoders </term> on a <term> Chinese-to-English translation task </term> . Our results show that <term> MBR decoding </term> can be used to tune <term> statistical MT </term> performance for specific <term> loss functions </term> .
<term> Topic signatures </term> can be useful in a number of <term> Natural Language Processing ( NLP ) </term> applications , such as <term> Word Sense Disambiguation ( WSD ) </term> and <term> Text Summarisation </term> . Our method takes advantage of the different way in which <term> word senses </term> are lexicalised in <term> English </term> and <term> Chinese </term> , and also exploits the large amount of <term> Chinese text </term> available in <term> corpora </term> and on the <term> Web </term> .
We also introduce a novel <term> classification method </term> based on <term> PER </term> which leverages <term> part of speech information </term> of the <term> words </term> contributing to the <term> word matches and non-matches </term> in the <term> sentence </term> . Our results show that <term> MT evaluation techniques </term> are able to produce useful <term> features </term> for <term> paraphrase classification </term> and to a lesser extent <term> entailment </term> .
Our results show that <term> MT evaluation techniques </term> are able to produce useful <term> features </term> for <term> paraphrase classification </term> and to a lesser extent <term> entailment </term> . Our <term> technique </term> gives a substantial improvement in <term> paraphrase classification accuracy </term> over all of the other <term> models </term> used in the experiments .
We use a <term> maximum likelihood criterion </term> to train a <term> log-linear block bigram model </term> which uses <term> real-valued features </term> ( e.g. a <term> language model score </term> ) as well as <term> binary features </term> based on the <term> block </term> identities themselves , e.g. block bigram features . Our <term> training algorithm </term> can easily handle millions of <term> features </term> .
Over the last decade , a variety of <term> SMT algorithms </term> have been built and empirically tested whereas little is known about the <term> computational complexity </term> of some of the fundamental problems of <term> SMT </term> . Our work aims at providing useful insights into the the <term> computational complexity </term> of those problems .
hide detail