TY - GEN
T1 - Few-shot Text Classification with Saliency-equivalent Concatenation
AU - Lin, Ying Jia
AU - Chang, Yu Fang
AU - Kao, Hung Yu
AU - Wang, Hsin Yang
AU - Liu, Mu
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - In few-shot text classification, the lack of significant features limits models from generalizing to data not included in the training set. Data augmentation is a solution to the classification tasks; however, the standard augmentation methods in natural language processing are not feasible in few-shot learning. In this study, we explore data augmentation in few-shot text classification. We propose saliency-equivalent concatenation (SEC)1. The core concept of SEC is to append additional key information to an input sentence to help a model understand the sentence easier. In the proposed method, we first leverage a pre-trained language model to generate several novel sentences for each sample in datasets. Then we leave the most relevant one and concatenate it with the original sentence as additional information for each sample. Our experiments on the two fewshot text classification tasks verified that the proposed method can boost the performance of meta-learning models and outperform the previous unsupervised data augmentation methods.
AB - In few-shot text classification, the lack of significant features limits models from generalizing to data not included in the training set. Data augmentation is a solution to the classification tasks; however, the standard augmentation methods in natural language processing are not feasible in few-shot learning. In this study, we explore data augmentation in few-shot text classification. We propose saliency-equivalent concatenation (SEC)1. The core concept of SEC is to append additional key information to an input sentence to help a model understand the sentence easier. In the proposed method, we first leverage a pre-trained language model to generate several novel sentences for each sample in datasets. Then we leave the most relevant one and concatenate it with the original sentence as additional information for each sample. Our experiments on the two fewshot text classification tasks verified that the proposed method can boost the performance of meta-learning models and outperform the previous unsupervised data augmentation methods.
UR - http://www.scopus.com/inward/record.url?scp=85143085035&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85143085035&partnerID=8YFLogxK
U2 - 10.1109/AIKE55402.2022.00019
DO - 10.1109/AIKE55402.2022.00019
M3 - Conference contribution
AN - SCOPUS:85143085035
T3 - Proceedings - 2022 IEEE 5th International Conference on Artificial Intelligence and Knowledge Engineering, AIKE 2022
SP - 74
EP - 81
BT - Proceedings - 2022 IEEE 5th International Conference on Artificial Intelligence and Knowledge Engineering, AIKE 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 5th IEEE International Conference on Artificial Intelligence and Knowledge Engineering, AIKE 2022
Y2 - 19 September 2022 through 21 September 2022
ER -