TY - GEN
T1 - Self-Guided Curriculum Learning for Neural Machine Translation
AU - Zhou, Lei
AU - Ding, Liang
AU - Duh, Kevin
AU - Watanabe, Shinji
AU - Sasano, Ryohei
AU - Takeda, Koichi
N1 - Funding Information:
Lei Zhou is supported by a project commissioned by the New Energy and Industrial Technology Development Organization (NEDO). The authors wish to thank the anonymous reviewers for their insightful comments and suggestions.
Publisher Copyright:
© 2021 Association for Computational Linguistics.
PY - 2021
Y1 - 2021
N2 - In supervised learning, a well-trained model should be able to recover ground truth accurately, i.e. the predicted labels are expected to resemble the ground truth labels as much as possible. Inspired by this, we formulate a difficulty criterion based on the recovery degrees of training examples. Motivated by the intuition that after skimming through the training corpus, the neural machine translation (NMT) model “knows” how to schedule a suitable curriculum according to learning difficulty, we propose a self-guided curriculum learning strategy that encourages the NMT model to learn from easy to hard on the basis of recovery degrees. Specifically, we adopt sentence-level BLEU score as the proxy of recovery degree. Experimental results on translation benchmarks including WMT14 English?German and WMT17 Chinese?English demonstrate that our proposed method considerably improves the recovery degree, thus consistently improving the translation performance.
AB - In supervised learning, a well-trained model should be able to recover ground truth accurately, i.e. the predicted labels are expected to resemble the ground truth labels as much as possible. Inspired by this, we formulate a difficulty criterion based on the recovery degrees of training examples. Motivated by the intuition that after skimming through the training corpus, the neural machine translation (NMT) model “knows” how to schedule a suitable curriculum according to learning difficulty, we propose a self-guided curriculum learning strategy that encourages the NMT model to learn from easy to hard on the basis of recovery degrees. Specifically, we adopt sentence-level BLEU score as the proxy of recovery degree. Experimental results on translation benchmarks including WMT14 English?German and WMT17 Chinese?English demonstrate that our proposed method considerably improves the recovery degree, thus consistently improving the translation performance.
UR - http://www.scopus.com/inward/record.url?scp=85112544154&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85112544154&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85112544154
T3 - IWSLT 2021 - 18th International Conference on Spoken Language Translation, Proceedings
SP - 206
EP - 214
BT - IWSLT 2021 - 18th International Conference on Spoken Language Translation, Proceedings
PB - Association for Computational Linguistics (ACL)
T2 - 18th International Conference on Spoken Language Translation, IWSLT 2021
Y2 - 5 August 2021 through 6 August 2021
ER -