TY - JOUR
T1 - Masked prompt learning for formal analogies beyond words
AU - Wang, Liyan
AU - Lepage, Yves
N1 - Funding Information:
The work is supported by China Scholarship Council (CSC) under the CSC Grant No. 202008050136.
Publisher Copyright:
© 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
PY - 2022
Y1 - 2022
N2 - Prompt learning, a recent thread in few-shot learning for pre-trained language models (PLMs), has been explored for completing word analogies in the extractive way. In this paper, we reformulate the analogy task as masked analogy completion task with the use of prompting to derive a generative model for analogies beyond words. We introduce a simple prompt-based fine-tuning paradigm for language modeling on answered prompts of analogies in the sequence-to-sequence framework. To convert discrete terms of analogies into linear sequences, we present a symbolic prompt template. The sequence-to-sequence model is fine-tuned to fill in the missing span of masked prompts deduced from different masking schemes on phrase analogies extracted from a small corpus. We analyze the out-of-distribution performance on sentence analogies which are unseen cases. Our experiments demonstrate that prompt-based fine-tuning with the objective of language modeling enables models to achieve significantly better performance on in-distribution cases than PLMs. Masked prompt learning with one-term masking exhibits the best out-of-distribution generalization on sentence analogies, with a difference of only 3 characters from references.
AB - Prompt learning, a recent thread in few-shot learning for pre-trained language models (PLMs), has been explored for completing word analogies in the extractive way. In this paper, we reformulate the analogy task as masked analogy completion task with the use of prompting to derive a generative model for analogies beyond words. We introduce a simple prompt-based fine-tuning paradigm for language modeling on answered prompts of analogies in the sequence-to-sequence framework. To convert discrete terms of analogies into linear sequences, we present a symbolic prompt template. The sequence-to-sequence model is fine-tuned to fill in the missing span of masked prompts deduced from different masking schemes on phrase analogies extracted from a small corpus. We analyze the out-of-distribution performance on sentence analogies which are unseen cases. Our experiments demonstrate that prompt-based fine-tuning with the objective of language modeling enables models to achieve significantly better performance on in-distribution cases than PLMs. Masked prompt learning with one-term masking exhibits the best out-of-distribution generalization on sentence analogies, with a difference of only 3 characters from references.
KW - Prompt learning
KW - analogies beyond words
KW - fine-tuning
KW - masked analogy completion
UR - http://www.scopus.com/inward/record.url?scp=85136233262&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85136233262&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85136233262
SN - 1613-0073
VL - 3174
SP - 1
EP - 14
JO - CEUR Workshop Proceedings
JF - CEUR Workshop Proceedings
T2 - 1st Workshop on the Interactions between Analogical Reasoning and Machine Learning at 31st International Joint Conference on Artificial Intelligence - 25th European Conference on Artificial Intelligence, IARML@IJCAI-ECAI 2022
Y2 - 23 July 2022
ER -