A92-1018 |
part-of-speech tagger based on a hidden
|
Markov model
|
. The methodology enables robust
|
A00-2020 |
Weischedel et al. ( 1993 ) applied
|
Markov Models
|
to tagging . Abney et al. ( 1999
|
A88-1005 |
a second -- order ( trigram )
|
Markov model
|
is uniquely determined by the
|
A00-1031 |
Underlying Model TnT uses second order
|
Markov models
|
for part-ofspeech tagging . The
|
A88-1028 |
statistical classifier based on Hidden
|
Markov Models
|
( HMM ) was developed for several
|
A88-1005 |
Moreover , the fourth -- order
|
Markov model
|
for the abstracted Thackeray
|
A00-2024 |
phrases come from ? We used a Hidden
|
Markov Model
|
( Baum , 1972 ) solution to the
|
A00-2024 |
heuristic rules to create a Hidden
|
Markov Model
|
. The Viterbi algorithm ( Viterbi
|
A00-1031 |
have shown that a tagger based on
|
Markov models
|
yields state-of-the-art results
|
A88-1005 |
than a full -- blown second order
|
Markov model
|
. Each state in a second -- order
|
A00-1031 |
• • I \ n ) For the
|
Markov model
|
, we need the inverse conditional
|
A00-1031 |
literature . For example , the
|
Markov model
|
tagger used in the comparison
|
A00-1044 |
First and foremost , the hidden
|
Markov model
|
is quite robust in the face of
|
A88-1028 |
technique based on the use of Hidden
|
Markov Models
|
( HMM ) was used as a language
|
A88-1005 |
however , neither a full -- blown
|
Markov model
|
using total vocabulary nor an
|
A00-2029 |
is a speaker-independent hidden
|
Markov model
|
system with context-dependent
|
A92-1018 |
interpre - tation . A form of
|
Markov model
|
has been widely used that assumes
|
A00-1032 |
a rule-base model or a hidden
|
Markov model
|
( HMM ) ( Manning and Schiitze
|
A88-1005 |
for any but the simplest order
|
Markov models
|
( orders zero and one ) , the
|
A00-2035 |
based on a combination of Hidden
|
Markov Models
|
( HMM ) and Maximum Entropy (
|