For Full-Text PDF, please login, if you are a member of IEICE,|
or go to Pay Per View on menu list, if you are a nonmember of IEICE.
Continuous Speech Segmentation Based on a Self-Learning Neuro-Fuzzy System
Ching-Tang HSIEH Mu-Chun SU Chih-Hsu HSU
IEICE TRANSACTIONS on Fundamentals of Electronics, Communications and Computer Sciences
Publication Date: 1996/08/25
Print ISSN: 0916-8508
Type of Manuscript: Special Section PAPER (Special Section on Digital Signal Processing)
speech segmentation, neural network, fuzzy systems, membership function,
Full Text: PDF(650.8KB)>>
For reducing requirement of large memory and minimizing computation complexity in a large-vocabulary continuous speech recognition system, speech segmentation plays an important role in speech recognition systems. In this paper, we formulate the speech segmentation as a two-phase problem. Phase 1 (frame labeling) involves labeling frames of speech data. Frames are classified into three types: (1) silence, (2) consonant and (3) vowel according to two segmentation features. In phase 2 (syllabic unit segmentation) we apply the concept of transition states to segment continuous speech data into syllabic units based on the labeled frames. The novel class of hyperrectangular composite neural networks (HRCNNs) is used to cluster frames. The HRCNNs integrate the rule-based approach and neural network paradigms, therefore, this special hybrid system may neutralize the disadvantages of each alternative. The parameters of the trained HRCNNs are utilized to extract both crisp and fuzzy classification rules. In our experiments, a database containing continuous reading-rate Mandarin speech recorded from newscast was utilized to illustrate the performance of the proposed speaker independent speech segmentation system. The effectiveness of the proposed segmentation system is confirmed by the experimental results.