TY - GEN
T1 - Mask Consistency and Contrast Regularization for Prompt-Based Learning
AU - Chien, Jen Tzung
AU - Chen, Chien Ching
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - This paper presents a new contrastive semi-supervised learning approach to data efficient language model where the fine-tuned mask language model is constructed with hard prompts for sentence representation. Importantly, the prompt-based pseudo labeling is merged in a mask language model (MLM) with data augmentation through a contrast loss which is utilized to pull together those similar samples and push apart the samples which do not belong to their augmentations. Furthermore, the mask consistency training is implemented for fine-tuned MLM to pursue the closeness between word prediction from weak and strong augmentations. Accordingly, this study develops a data efficient solution which improves the model generalization and leverages the rich information from unlabeled data for few-shot text classification. The experiments on natural language understanding in few-shot and semi-supervised settings show that the proposed method considerably improves the performance by using the contrastive prompt-based learning and the mask consistency training.
AB - This paper presents a new contrastive semi-supervised learning approach to data efficient language model where the fine-tuned mask language model is constructed with hard prompts for sentence representation. Importantly, the prompt-based pseudo labeling is merged in a mask language model (MLM) with data augmentation through a contrast loss which is utilized to pull together those similar samples and push apart the samples which do not belong to their augmentations. Furthermore, the mask consistency training is implemented for fine-tuned MLM to pursue the closeness between word prediction from weak and strong augmentations. Accordingly, this study develops a data efficient solution which improves the model generalization and leverages the rich information from unlabeled data for few-shot text classification. The experiments on natural language understanding in few-shot and semi-supervised settings show that the proposed method considerably improves the performance by using the contrastive prompt-based learning and the mask consistency training.
KW - Semi-supervised learning
KW - contrastive learning
KW - prompt-based learning
KW - pseudo labeling
UR - http://www.scopus.com/inward/record.url?scp=85205027397&partnerID=8YFLogxK
U2 - 10.1109/IJCNN60899.2024.10650956
DO - 10.1109/IJCNN60899.2024.10650956
M3 - Conference contribution
AN - SCOPUS:85205027397
T3 - Proceedings of the International Joint Conference on Neural Networks
BT - 2024 International Joint Conference on Neural Networks, IJCNN 2024 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2024 International Joint Conference on Neural Networks, IJCNN 2024
Y2 - 30 June 2024 through 5 July 2024
ER -