TY - GEN
T1 - Exploiting Paraphrasers and Inverse Paraphrasers
T2 - 7th International Conference on Computer Science and Artificial Intelligence, CSAI 2023
AU - Du, Zhendong
AU - Hashimoto, Kenji
N1 - Publisher Copyright:
© 2023 ACM.
PY - 2023/12/8
Y1 - 2023/12/8
N2 - In the realm of enhancing English writing fluency, the scarcity of high-quality training data has perennially posed a significant challenge. Moreover, elevating the fluency of writing while ensuring the preservation of semantic integrity compounds the intricacies of this task. In this study, we introduce and implement a style converter rooted in the Paraphraser and Inverse Paraphraser methodologies, aimed at ameliorating English writing fluency. Concurrently, this converter facilitated the generation of a voluminous corpus of synthetic training data. Utilizing this data, we fine-tuned GPT-2 to forge an English text style transfer model. Remarkably, despite our model being trained on a dataset substantially smaller than that of prevailing baseline methods, it exhibited exemplary performance across multiple evaluation metrics, even surpassing these baselines on certain pivotal indices. These findings corroborate the efficacy of our approach and underscore its immense potential in the domain of English writing fluency enhancement. This investigation not only offers a novel optimization strategy for English composition but also furnishes researchers in cognate fields with fresh research perspectives and methodologies.
AB - In the realm of enhancing English writing fluency, the scarcity of high-quality training data has perennially posed a significant challenge. Moreover, elevating the fluency of writing while ensuring the preservation of semantic integrity compounds the intricacies of this task. In this study, we introduce and implement a style converter rooted in the Paraphraser and Inverse Paraphraser methodologies, aimed at ameliorating English writing fluency. Concurrently, this converter facilitated the generation of a voluminous corpus of synthetic training data. Utilizing this data, we fine-tuned GPT-2 to forge an English text style transfer model. Remarkably, despite our model being trained on a dataset substantially smaller than that of prevailing baseline methods, it exhibited exemplary performance across multiple evaluation metrics, even surpassing these baselines on certain pivotal indices. These findings corroborate the efficacy of our approach and underscore its immense potential in the domain of English writing fluency enhancement. This investigation not only offers a novel optimization strategy for English composition but also furnishes researchers in cognate fields with fresh research perspectives and methodologies.
KW - Data augmentation
KW - English writing assistant
KW - Paraphrase Generation
KW - Style transfer
UR - http://www.scopus.com/inward/record.url?scp=85188289599&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85188289599&partnerID=8YFLogxK
U2 - 10.1145/3638584.3638618
DO - 10.1145/3638584.3638618
M3 - Conference contribution
AN - SCOPUS:85188289599
T3 - ACM International Conference Proceeding Series
SP - 346
EP - 352
BT - CSAI 2023 - 2023 7th International Conference on Computer Science and Artificial Intelligence
PB - Association for Computing Machinery
Y2 - 8 December 2023 through 10 December 2023
ER -