other,8-1-H01-1017,bq To support engaging human users in robust , <term> mixed-initiative speech dialogue interactions </term> which reach beyond current capabilities in <term> dialogue systems </term> , the <term> DARPA Communicator program </term> [ 1 ] is funding the development of a <term> distributed message-passing infrastructure </term> for <term> dialogue systems </term> which all <term> Communicator </term> participants are using .
tech,4-2-H01-1055,bq However , the improved <term> speech recognition </term> has brought to light a new problem : as <term> dialog systems </term> understand more of what the <term> user </term> tells them , they need to be more sophisticated at responding to the <term> user </term> .
other,26-1-N01-1003,bq <term> Sentence planning </term> is a set of inter-related but distinct tasks , one of which is <term> sentence scoping </term> , i.e. the choice of <term> syntactic structure </term> for elementary <term> speech acts </term> and the decision of how to combine them into one or more <term> sentences </term> .
other,10-2-N03-1012,bq We apply our <term> system </term> to the task of <term> scoring </term> alternative <term> speech recognition hypotheses ( SRH ) </term> in terms of their <term> semantic coherence </term> .
other,18-3-N03-1012,bq We conducted an <term> annotation experiment </term> and showed that <term> human annotators </term> can reliably differentiate between semantically coherent and incoherent <term> speech recognition hypotheses </term> .
other,9-1-N03-2003,bq Sources of <term> training data </term> suitable for <term> language modeling </term> of <term> conversational speech </term> are limited .
tech,15-3-P03-1030,bq Motivated by these arguments , we introduce a number of new performance enhancing techniques including <term> part of speech tagging </term> , new <term> similarity measures </term> and expanded <term> stop lists </term> .
tech,19-3-P03-1031,bq Since multiple <term> candidates </term> for the <term> understanding </term> result can be obtained for a <term> user utterance </term> due to the <term> ambiguity </term> of <term> speech understanding </term> , it is not appropriate to decide on a single <term> understandingresult </term> after each <term> user utterance </term> .
other,8-1-C04-1103,bq <term> Machine transliteration/back-transliteration </term> plays an important role in many <term> multilingual speech and language applications </term> .
other,12-3-I05-5003,bq We also introduce a novel <term> classification method </term> based on <term> PER </term> which leverages <term> part of speech information </term> of the <term> words </term> contributing to the <term> word matches and non-matches </term> in the <term> sentence </term> .
tech,36-12-J05-1003,bq Although the experiments in this article are on <term> natural language parsing ( NLP ) </term> , the <term> approach </term> should be applicable to many other <term> NLP problems </term> which are naturally framed as <term> ranking tasks </term> , for example , <term> speech recognition </term> , <term> machine translation </term> , or <term> natural language generation </term> .
other,70-5-E06-1035,bq Examination of the effect of <term> features </term> shows that <term> predicting top-level and predicting subtopic boundaries </term> are two distinct tasks : ( 1 ) for predicting <term> subtopic boundaries </term> , the <term> lexical cohesion-based approach </term> alone can achieve competitive results , ( 2 ) for <term> predicting top-level boundaries </term> , the <term> machine learning approach </term> that combines <term> lexical-cohesion and conversational features </term> performs best , and ( 3 ) <term> conversational cues </term> , such as <term> cue phrases </term> and <term> overlapping speech </term> , are better indicators for the top-level prediction task .
tech,16-2-C90-3014,bq The approach of <term> KPSG provides </term> an explicit development model for constructing a computational <term> phonological system </term> : <term> speech recognition </term> and <term> synthesis system </term> .
other,7-1-H90-1060,bq This paper reports on two contributions to <term> large vocabulary continuous speech recognition </term> .
lr,27-2-H90-1060,bq First , we present a new paradigm for <term> speaker-independent ( SI ) training </term> of <term> hidden Markov models ( HMM ) </term> , which uses a large amount of <term> speech </term> from a few <term> speakers </term> instead of the traditional practice of using a little <term> speech </term> from many <term> speakers </term> .
lr,41-2-H90-1060,bq First , we present a new paradigm for <term> speaker-independent ( SI ) training </term> of <term> hidden Markov models ( HMM ) </term> , which uses a large amount of <term> speech </term> from a few <term> speakers </term> instead of the traditional practice of using a little <term> speech </term> from many <term> speakers </term> .
lr,27-3-H90-1060,bq In addition , combination of the <term> training speakers </term> is done by averaging the <term> statistics > </term> of <term> independently trained models </term> rather than the usual pooling of all the <term> speech data </term> from many <term> speakers </term> prior to <term> training </term> .
lr,23-6-H90-1060,bq Second , we show a significant improvement for <term> speaker adaptation ( SA ) </term> using the new <term> SI corpus </term> and a small amount of <term> speech </term> from the new ( target ) <term> speaker </term> .
tech,5-1-C92-3165,bq This paper introduces a robust <term> interactive method for speech understanding </term> .
other,35-3-H92-1003,bq We summarize the motivation for this effort , the goals , the implementation of a <term> multi-site data collection paradigm </term> , and the accomplishments of <term> MADCOW </term> in monitoring the <term> collection </term> and distribution of 12,000 <term> utterances </term> of <term> spontaneous speech </term> from five sites for use in a <term> multi-site common evaluation of speech , natural language and spoken language
hide detail