The search functionality is under construction.

IEICE TRANSACTIONS on Fundamentals

Continuous Speech Segmentation Based on a Self-Learning Neuro-Fuzzy System

Ching-Tang HSIEH, Mu-Chun SU, Chih-Hsu HSU

  • Full Text Views

    0

  • Cite this

Summary :

For reducing requirement of large memory and minimizing computation complexity in a large-vocabulary continuous speech recognition system, speech segmentation plays an important role in speech recognition systems. In this paper, we formulate the speech segmentation as a two-phase problem. Phase 1 (frame labeling) involves labeling frames of speech data. Frames are classified into three types: (1) silence, (2) consonant and (3) vowel according to two segmentation features. In phase 2 (syllabic unit segmentation) we apply the concept of transition states to segment continuous speech data into syllabic units based on the labeled frames. The novel class of hyperrectangular composite neural networks (HRCNNs) is used to cluster frames. The HRCNNs integrate the rule-based approach and neural network paradigms, therefore, this special hybrid system may neutralize the disadvantages of each alternative. The parameters of the trained HRCNNs are utilized to extract both crisp and fuzzy classification rules. In our experiments, a database containing continuous reading-rate Mandarin speech recorded from newscast was utilized to illustrate the performance of the proposed speaker independent speech segmentation system. The effectiveness of the proposed segmentation system is confirmed by the experimental results.

Publication
IEICE TRANSACTIONS on Fundamentals Vol.E79-A No.8 pp.1180-1187
Publication Date
1996/08/25
Publicized
Online ISSN
DOI
Type of Manuscript
Special Section PAPER (Special Section on Digital Signal Processing)
Category

Authors

Keyword