We present a method that determines articulatory movements from speech acoustics using an HMM (Hidden Markov Model)-based speech production model. The model statistically generates speech acoustics and articulatory movements from a given phonemic string. It consists of HMMs of articulatory movements for each phoneme and an articulatory-to-acoustic mapping. For a given speech acoustics, maximum likelihood estimation of the articulatory parameters is the statistical model is presented. An experiment was conducted to evaluate the method on sentences, and estimated articulatory parameters were compared with the observed parameters. The rms error of the estimated articulatory parameters is about 2.12 mm on average.
|ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
|Published - 2002
|2002 IEEE International Conference on Acustics, Speech, and Signal Processing - Orlando, FL, United States
Duration: 2002 May 13 → 2002 May 17
ASJC Scopus subject areas
- Signal Processing
- Electrical and Electronic Engineering