TY - JOUR
T1 - Statistical language modeling with a class-based n-multigram model
AU - Deligne, Sabine
AU - Sagisaka, Yoshinori
N1 - Copyright:
Copyright 2017 Elsevier B.V., All rights reserved.
PY - 2000/7
Y1 - 2000/7
N2 - In this paper, we present a stochastic language-modeling tool which aims at retrieving variable-length phrases (multigrams), assuming n-gram dependencies between them, hence the name of the model: n-multigram. The estimation of the probability distribution of the phrases is intermixed with a phrase-clustering procedure in a way which jointly optimizes the likelihood of the data. As a result, the language data are iteratively structured at both a paradigmatic and a syntagmatic level in a fully integrated way. We evaluate the 2-multigram model as a statistical language model on ATIS, a task-oriented database consisting of air travel reservations. Experiments show that the 2-multigrarn model allows a reduction of 10% of the word error rate on ATIS with respect to the usual trigram model, with 25% fewer parameters than in the trigram model. In addition, we illustrate the ability of this model to merge semantically related phrases of different lengths into a common class.
AB - In this paper, we present a stochastic language-modeling tool which aims at retrieving variable-length phrases (multigrams), assuming n-gram dependencies between them, hence the name of the model: n-multigram. The estimation of the probability distribution of the phrases is intermixed with a phrase-clustering procedure in a way which jointly optimizes the likelihood of the data. As a result, the language data are iteratively structured at both a paradigmatic and a syntagmatic level in a fully integrated way. We evaluate the 2-multigram model as a statistical language model on ATIS, a task-oriented database consisting of air travel reservations. Experiments show that the 2-multigrarn model allows a reduction of 10% of the word error rate on ATIS with respect to the usual trigram model, with 25% fewer parameters than in the trigram model. In addition, we illustrate the ability of this model to merge semantically related phrases of different lengths into a common class.
UR - http://www.scopus.com/inward/record.url?scp=0034230088&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=0034230088&partnerID=8YFLogxK
U2 - 10.1006/csla.2000.0146
DO - 10.1006/csla.2000.0146
M3 - Article
AN - SCOPUS:0034230088
SN - 0885-2308
VL - 14
SP - 261
EP - 279
JO - Computer Speech and Language
JF - Computer Speech and Language
IS - 3
ER -