TY - JOUR
T1 - Improving speech understanding accuracy with limited training data using multiple language models and multiple understanding models
AU - Katsumaru, Masaki
AU - Nakano, Mikio
AU - Komatani, Kazunori
AU - Funakoshi, Kotaro
AU - Ogata, Tetsuya
AU - Okuno, Hiroshi G.
PY - 2009
Y1 - 2009
N2 - We aim to improve a speech understanding module with a small amount of training data. A speech understanding module uses a language model (LM) and a language understanding model (LUM). A lot of training data are needed to improve the models. Such data collection is, however, difficult in an actual process of development. We therefore design and develop a new framework that uses multiple LMs and LUMs to improve speech understanding accuracy under various amounts of training data. Even if the amount of available training data is small, each LM and each LUM can deal well with different types of utterances and more utterances are understood by using multiple LM and LUM. As one implementation of the framework, we develop a method for selecting the most appropriate speech understanding result from several candidates. The selection is based on probabilities of correctness calculated by logistic regressions. We evaluate our framework with various amounts of training data.
AB - We aim to improve a speech understanding module with a small amount of training data. A speech understanding module uses a language model (LM) and a language understanding model (LUM). A lot of training data are needed to improve the models. Such data collection is, however, difficult in an actual process of development. We therefore design and develop a new framework that uses multiple LMs and LUMs to improve speech understanding accuracy under various amounts of training data. Even if the amount of available training data is small, each LM and each LUM can deal well with different types of utterances and more utterances are understood by using multiple LM and LUM. As one implementation of the framework, we develop a method for selecting the most appropriate speech understanding result from several candidates. The selection is based on probabilities of correctness calculated by logistic regressions. We evaluate our framework with various amounts of training data.
KW - Limited training data
KW - Multiple language models and language understanding models
KW - Speech understanding
UR - http://www.scopus.com/inward/record.url?scp=70450218193&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=70450218193&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:70450218193
SN - 2308-457X
SP - 2735
EP - 2738
JO - Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
JF - Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
T2 - 10th Annual Conference of the International Speech Communication Association, INTERSPEECH 2009
Y2 - 6 September 2009 through 10 September 2009
ER -