In this paper we show how two standard outputs from <term> information extraction ( IE ) systems </term> - <term> named entity annotations </term> and <term> scenario templates </term> - can be used to enhance access to <term> text collections </term> via a standard <term> text browser </term> . We describe how this information is used in a <term> prototype system </term> designed to support <term> information workers </term> ' access to a <term> pharmaceutical news archive </term> as part of their <term> industry watch </term> function .
We describe how this information is used in a <term> prototype system </term> designed to support <term> information workers </term> ' access to a <term> pharmaceutical news archive </term> as part of their <term> industry watch </term> function . We also report results of a preliminary , <term> qualitative user evaluation </term> of the <term> system </term> , which while broadly positive indicates further work needs to be done on the <term> interface </term> to make <term> users </term> aware of the increased potential of <term> IE-enhanced text browsers </term> .
The purpose of this research is to test the efficacy of applying <term> automated evaluation techniques </term> , originally devised for the <term> evaluation </term> of <term> human language learners </term> , to the <term> output </term> of <term> machine translation ( MT ) systems </term> . We believe that these <term> evaluation techniques </term> will provide information about both the <term> human language learning process </term> , the <term> translation process </term> and the <term> development </term> of <term> machine translation systems </term> .
Even more illuminating was the factors on which the <term> assessors </term> made their decisions . We tested this to see if similar criteria could be elicited from duplicating the experiment using <term> machine translation output </term> .
<term> Listen-Communicate-Show ( LCS ) </term> is a new paradigm for <term> human interaction with data sources </term> . We integrate a <term> spoken language understanding system </term> with <term> intelligent mobile agents </term> that mediate between <term> users </term> and <term> information sources </term> .
We integrate a <term> spoken language understanding system </term> with <term> intelligent mobile agents </term> that mediate between <term> users </term> and <term> information sources </term> . We have built and will demonstrate an application of this approach called <term> LCS-Marine </term> .
<term> Requestors </term> can also instruct the <term> system </term> to notify them when the status of a <term> request </term> changes or when a <term> request </term> is complete . We have demonstrated this capability in several field exercises with the Marines and are currently developing applications of this <term> technology </term> in <term> new domains </term> .
The issue of <term> system response </term> to <term> users </term> has been extensively studied by the <term> natural language generation community </term> , though rarely in the context of <term> dialog systems </term> . We show how research in <term> generation </term> can be adapted to <term> dialog systems </term> , and how the high cost of hand-crafting <term> knowledge-based generation systems </term> can be overcome by employing <term> machine learning techniques </term> .
In this paper , we address the problem of combining several <term> language models ( LMs ) </term> . We find that simple <term> interpolation methods </term> , like <term> log-linear and linear interpolation </term> , improve the <term> performance </term> but fall short of the <term> performance </term> of an <term> oracle </term> .
Actually , the <term> oracle </term> acts like a <term> dynamic combiner </term> with <term> hard decisions </term> using the <term> reference </term> . We provide experimental results that clearly show the need for a <term> dynamic language model combination </term> to improve the <term> performance </term> further .
We provide experimental results that clearly show the need for a <term> dynamic language model combination </term> to improve the <term> performance </term> further . We suggest a method that mimics the behavior of the <term> oracle </term> using a <term> neural network </term> or a <term> decision tree </term> .
The method amounts to tagging <term> LMs </term> with <term> confidence measures </term> and picking the best <term> hypothesis </term> corresponding to the <term> LM </term> with the best <term> confidence </term> . We describe a three-tiered approach for <term> evaluation </term> of <term> spoken dialogue systems </term> .
The three tiers measure <term> user satisfaction </term> , <term> system support of mission success </term> and <term> component performance </term> . We describe our use of this approach in numerous fielded <term> user studies </term> conducted with the U.S. military .
In this paper , we present <term> SPoT </term> , a <term> sentence planner </term> , and a new methodology for automatically training <term> SPoT </term> on the basis of <term> feedback </term> provided by <term> human judges </term> . We reconceptualize the task into two distinct phases .
The <term> SPR </term> uses <term> ranking rules </term> automatically learned from <term> training data </term> . We show that the trained <term> SPR </term> learns to select a <term> sentence plan </term> whose <term> rating </term> on average is only 5 % worse than the <term> top human-ranked sentence plan </term> .
In this paper , we compare the relative effects of <term> segment order </term> , <term> segmentation </term> and <term> segment contiguity </term> on the <term> retrieval performance </term> of a <term> translation memory system </term> . We take a selection of both <term> bag-of-words and segment order-sensitive string comparison methods </term> , and run each over both <term> character - and word-segmented data </term> , in combination with a range of <term> local segment contiguity models </term> ( in the form of <term> N-grams </term> ) .
Further , in their optimum <term> configuration </term> , <term> bag-of-words methods </term> are shown to be equivalent to <term> segment order-sensitive methods </term> in terms of <term> retrieval accuracy </term> , but much faster . We also provide evidence that our findings are scalable .
While <term> paraphrasing </term> is critical both for <term> interpretation and generation of natural language </term> , current systems use manual or semi-automatic methods to collect <term> paraphrases </term> . We present an <term> unsupervised learning algorithm </term> for <term> identification of paraphrases </term> from a <term> corpus of multiple English translations </term> of the same <term> source text </term> .
The value of this approach is that as the <term> operational semantics </term> of <term> natural language applications </term> improve , even larger improvements are possible . We provide a <term> logical definition </term> of <term> Minimalist grammars </term> , that are <term> Stabler 's formalization </term> of <term> Chomsky 's minimalist program </term> .
In this paper We experimentally evaluate a <term> trainable sentence planner </term> for a <term> spoken dialogue system </term> by eliciting <term> subjective human judgments </term> .
hide detail