TY - GEN
T1 - ABCD
T2 - 10th International Conference on Theory and Practice of Natural Computing, TPNC 2021
AU - Yang, Shuo
AU - Lepage, Yves
N1 - Publisher Copyright:
© 2021, Springer Nature Switzerland AG.
PY - 2021
Y1 - 2021
N2 - We propose an analogy-based data augmentation approach for sentiment and style transfer named Analogy-Based Controllable Data Augmentation (ABCD). The object of data augmentation is to expand the number of sentences based on a limited amount of available data. We are given two unpaired corpora with different styles. In data augmentation, we retain the original text style while changing words to generate new sentences. We first train a self-attention-based convolutional neural network to compute the distribution of the contribution of each word to style in a given sentence. We call the words with high style contribution style-characteristic words. By substituting content words and style-characteristic words separately, we generate two new sentences. We use an analogy between the original sentence and these two additional sentences to generate another sentence. The results show that our proposed approach decrease perplexity by about 4 points and outperforms baselines on three transfer datasets.
AB - We propose an analogy-based data augmentation approach for sentiment and style transfer named Analogy-Based Controllable Data Augmentation (ABCD). The object of data augmentation is to expand the number of sentences based on a limited amount of available data. We are given two unpaired corpora with different styles. In data augmentation, we retain the original text style while changing words to generate new sentences. We first train a self-attention-based convolutional neural network to compute the distribution of the contribution of each word to style in a given sentence. We call the words with high style contribution style-characteristic words. By substituting content words and style-characteristic words separately, we generate two new sentences. We use an analogy between the original sentence and these two additional sentences to generate another sentence. The results show that our proposed approach decrease perplexity by about 4 points and outperforms baselines on three transfer datasets.
KW - Affective computing
KW - Computing with words
KW - Natural language processing
KW - Neural networks
UR - http://www.scopus.com/inward/record.url?scp=85119886583&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85119886583&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-90425-8_6
DO - 10.1007/978-3-030-90425-8_6
M3 - Conference contribution
AN - SCOPUS:85119886583
SN - 9783030904241
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 69
EP - 81
BT - Theory and Practice of Natural Computing - 10th International Conference, TPNC 2021, Proceedings
A2 - Aranha, Claus
A2 - Martín-Vide, Carlos
A2 - Vega-Rodríguez, Miguel A.
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 7 December 2021 through 10 December 2021
ER -