<term> Sentence planning </term> is a set of inter-related but distinct tasks , one of which is <term> sentence scoping </term> , i.e. the choice of <term> syntactic structure </term> for elementary <term> speech acts </term> and the decision of how to combine them into one or more <term> sentences </term> .
We reconceptualize the task into two distinct phases .
In particular , <term> range concatenation languages [ RCL ] </term> can be parsed in <term> polynomial time </term> and many classical <term> grammatical formalisms </term> can be translated into equivalent <term> RCGs </term> without increasing their <term> worst-case parsing time complexity </term> .
For example , after <term> translation </term> into an equivalent <term> RCG </term> , any <term> tree adjoining grammar </term> can be parsed in <term> O ( n6 ) time </term> .
In this paper , we introduce a <term> generative probabilistic optical character recognition ( OCR ) model </term> that describes an end-to-end process in the <term> noisy channel framework </term> , progressing from generation of <term> true text </term> through its transformation into the <term> noisy output </term> of an <term> OCR system </term> .
We create a <term> word-trie </term> , transform it into a <term> minimal DFA </term> , then identify <term> hubs </term> .
It gives users the ability to spend their time finding more data relevant to their task , and gives them translingual reach into other <term> languages </term> by leveraging <term> human language technology </term> .
A novel <term> evaluation scheme </term> is proposed which accounts for the effect of <term> polysemy </term> on the <term> clusters </term> , offering us a good insight into the potential and limitations of <term> semantically classifying </term><term> undisambiguated SCF data </term> .
We incorporate this analysis into a <term> diagnostic tool </term> intended for <term> developers </term> of <term> machine translation systems </term> , and demonstrate how our application can be used by <term> developers </term> to explore <term> patterns </term> in <term> machine translation output </term> .
Theory will be put into practice .
The strength of our <term> approach </term> is that it allows a <term> tree </term> to be represented as an arbitrary set of <term> features </term> , without concerns about how these <term> features </term> interact or overlap and without the need to define a <term> derivation </term> or a <term> generative model </term> which takes these <term> features </term> into account .
We define a <term> paraphrase probability </term> that allows <term> paraphrases </term> extracted from a <term> bilingual parallel corpus </term> to be ranked using <term> translation probabilities </term> , and show how it can be refined to take <term> contextual information </term> into account .
Our work aims at providing useful insights into the the <term> computational complexity </term> of those problems .
This paper describes <term> FERRET </term> , an <term> interactive question-answering ( Q/A ) system </term> designed to address the challenges of integrating <term> automatic Q/A </term> applications into real-world environments .
The <term> LOGON MT demonstrator </term> assembles independently valuable <term> general-purpose NLP components </term> into a <term> machine translation pipeline </term> that capitalizes on <term> output quality </term> .
With the aid of a <term> logic-based grammar formalism </term> called <term> extraposition grammars </term> , <term> Chat-80 </term> translates <term> English questions </term> into the <term> Prolog </term><term> subset of logic </term> .
The resulting <term> logical expression </term> is then transformed by a <term> planning algorithm </term> into efficient <term> Prolog </term> , cf. <term> query optimisation </term> in a <term> relational database </term> .
This method of using <term> expectations </term> to aid the understanding of <term> scruffy texts </term> has been incorporated into a working <term> computer program </term> called <term> NOMAD </term> , which understands <term> scruffy texts </term> in the domain of Navy messages .
The process of transforming a <term> disposition </term> into a <term> proposition </term> is referred to as <term> explicitation </term> or <term> restoration </term> .
The <term> linguistic structure </term> consists of segments of the <term> discourse </term> into which the <term> utterances </term> naturally aggregate .
hide detail