Hidden Markov model for acoustic syllabification and phoneme class labelling in continuous speech

Prinsloo G.J. ; Coetzer M.W. (1989)

Conference Paper

The use of a constrained hidden Markov model for the segmentation of syllables and syllable classes in continuous speech is described. The implementation involves a combination of phonological knowledge, speaker-adaptable features, vector quantization, and a hidden Markov modeling technique. Syllable classes are represented as hidden Markov model states, and phonotactic constraints are used to define state connections. A robust observation sequence is obtained from the decision planes of two single-layer neural networks, which are used to make voiced/unvoiced and (syllabic) nucleus/nonnucleus classifications, respectively. Speaker-independent tests, conducted on a broadcast quality database, revealed a syllable class recognition rate of 96.5% and a detection rate of 97.1% for syllables.

Please refer to this item in SUNScholar by using the following persistent URL: http://hdl.handle.net/10019.1/13209
This item appears in the following collections: