2025.naacl-srw.39@ACL

Total: 1

#1 Paraphrase-based Contrastive Learning for Sentence Pair Modeling [PDF] [Copy] [Kimi] [REL]

Authors: Seiji Sugiyama, Risa Kondo, Tomoyuki Kajiwara, Takashi Ninomiya

To improve the performance of sentence pair modeling tasks, we propose an additional pre-training method, also known as transfer fine-tuning, for pre-trained masked language models.Pre-training for masked language modeling is not necessarily designed to bring semantically similar sentences closer together in the embedding space.Our proposed method aims to improve the performance of sentence pair modeling by applying contrastive learning to pre-trained masked language models, in which sentence embeddings of paraphrase pairs are made similar to each other.While natural language inference corpora, which are standard in previous studies on contrastive learning, are not available on a large-scale for non-English languages, our method can construct a training corpus for contrastive learning from a raw corpus and a paraphrase dictionary at a low cost.Experimental results on four sentence pair modeling tasks revealed the effectiveness of our method in both English and Japanese.

Subject: NAACL.2025 - Student Research Workshop