|
For Full-Text PDF, please login, if you are a member of IEICE,
or go to Pay Per View on menu list, if you are a nonmember of IEICE.
|
Succeeding Word Prediction for Speech Recognition Based on Stochastic Language Model
Min ZHOU Seiichi NAKAGAWA
Publication
IEICE TRANSACTIONS on Information and Systems
Vol.E79-D
No.4
pp.333-342 Publication Date: 1996/04/25 Online ISSN:
DOI: Print ISSN: 0916-8532 Type of Manuscript: PAPER Category: Speech Processing and Acoustics Keyword: stochastic language model, bigram, trigram, HMM, bigram-HMM, context-free grammar, entropy, perplexity, prediction rate,
Full Text: PDF(783.5KB)>>
Summary:
For the purpose of automatic speech recognition, language models (LMs) are used to predict possible succeeding words for a given partial word sequence and thereby to reduce the search space. In this paper several kinds of stochastic language models (SLMs) are evaluated-bigram, trigram, hidden Markov model (HMM), bigram-HMM, stochastic context-free grammar (SCFG) and hand-written Bunsetsu Grammar. To compare the predictive power of these SLMs, the evaluation was conducted from two points of views: (1) relationship between the number of model parameters and entropy, (2) predictive rate of succeeding part of speech (POS) and succeeding word. We propose a new type of bigram-HMM and compare it with the other models. Two kinds of approximations are tried and examined through experiments. Results based on both of English Brown-Corpus and Japanese ATR dialog database showed that the extended bigram-HMM had better performance than the others and was more suitable to be a language model.
|
|