Advanced computational models and learning theories for spoken language processing

Atsushi Nakamura*, Shinji Watanabe, Takaaki Hori, Erik McDermott, Shigeru Katagiri

*Corresponding author for this work

Research output: Contribution to journalReview articlepeer-review

2 Citations (Scopus)

Abstract

Various methods for fast search through finite-state machines, Bayesian solutions for modeling and classification of speech, and a training approach for minimizing errors in large vocabulary continuous speech recognition (LVCSR) technology are discussed. The development of effective speech recognition decoders requires understanding software programming skills and sufficient understanding of LVCSR technology. The weighted finite-state transducer (WFST) framework provides an alternative to LVCSR and enables efficient global optimization of the search space and a one-pass decoding over the speech input using all knowledge simultaneously. The MCE-based training framework is extended to make full use of WFST to obtain directly model word accuracy.

Original languageEnglish
Pages (from-to)5-9+26
JournalIEEE Computational Intelligence Magazine
Volume1
Issue number2
Publication statusPublished - 2006 May
Externally publishedYes

ASJC Scopus subject areas

  • Theoretical Computer Science
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Advanced computational models and learning theories for spoken language processing'. Together they form a unique fingerprint.

Cite this