A Multi-level Supervised Contrastive Learning Framework for Low-Resource
Natural Language Inference
- URL: http://arxiv.org/abs/2205.15550v1
- Date: Tue, 31 May 2022 05:54:18 GMT
- Title: A Multi-level Supervised Contrastive Learning Framework for Low-Resource
Natural Language Inference
- Authors: Shu'ang Li, Xuming Hu, Li Lin, Aiwei Liu, Lijie Wen, Philip S. Yu
- Abstract summary: Natural Language Inference (NLI) is a growingly essential task in natural language understanding.
Here we propose a multi-level supervised contrastive learning framework named MultiSCL for low-resource natural language inference.
- Score: 54.678516076366506
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Natural Language Inference (NLI) is a growingly essential task in natural
language understanding, which requires inferring the relationship between the
sentence pairs (premise and hypothesis). Recently, low-resource natural
language inference has gained increasing attention, due to significant savings
in manual annotation costs and a better fit with real-world scenarios. Existing
works fail to characterize discriminative representations between different
classes with limited training data, which may cause faults in label prediction.
Here we propose a multi-level supervised contrastive learning framework named
MultiSCL for low-resource natural language inference. MultiSCL leverages a
sentence-level and pair-level contrastive learning objective to discriminate
between different classes of sentence pairs by bringing those in one class
together and pushing away those in different classes. MultiSCL adopts a data
augmentation module that generates different views for input samples to better
learn the latent representation. The pair-level representation is obtained from
a cross attention module. We conduct extensive experiments on two public NLI
datasets in low-resource settings, and the accuracy of MultiSCL exceeds other
models by 3.1% on average. Moreover, our method outperforms the previous
state-of-the-art method on cross-domain tasks of text classification.
Related papers
- Constrained Multi-Layer Contrastive Learning for Implicit Discourse Relationship Recognition [9.644808110078833]
We adapt a supervised contrastive learning (CL) method, label- and instance-centered CL, to enhance representation learning.
Experimental results on PDTB 2.0 and PDTB 3.0 show that our approach can significantly improve the performance on both multi-class classification and binary classification.
arXiv Detail & Related papers (2024-09-07T17:55:41Z) - Efficient Spoken Language Recognition via Multilabel Classification [53.662747523872305]
We show that our models obtain competitive results while being orders of magnitude smaller and faster than current state-of-the-art methods.
Our multilabel strategy is more robust to unseen non-target languages compared to multiclass classification.
arXiv Detail & Related papers (2023-06-02T23:04:19Z) - MaPLe: Multi-modal Prompt Learning [54.96069171726668]
We propose Multi-modal Prompt Learning (MaPLe) for both vision and language branches to improve alignment between the vision and language representations.
Compared with the state-of-the-art method Co-CoOp, MaPLe exhibits favorable performance and achieves an absolute gain of 3.45% on novel classes.
arXiv Detail & Related papers (2022-10-06T17:59:56Z) - Multi-level Contrastive Learning for Cross-lingual Spoken Language
Understanding [90.87454350016121]
We develop novel code-switching schemes to generate hard negative examples for contrastive learning at all levels.
We develop a label-aware joint model to leverage label semantics for cross-lingual knowledge transfer.
arXiv Detail & Related papers (2022-05-07T13:44:28Z) - Bridging the Gap between Language Models and Cross-Lingual Sequence
Labeling [101.74165219364264]
Large-scale cross-lingual pre-trained language models (xPLMs) have shown effectiveness in cross-lingual sequence labeling tasks.
Despite the great success, we draw an empirical observation that there is a training objective gap between pre-training and fine-tuning stages.
In this paper, we first design a pre-training task tailored for xSL named Cross-lingual Language Informative Span Masking (CLISM) to eliminate the objective gap.
Second, we present ContrAstive-Consistency Regularization (CACR), which utilizes contrastive learning to encourage the consistency between representations of input parallel
arXiv Detail & Related papers (2022-04-11T15:55:20Z) - Multi-Level Contrastive Learning for Cross-Lingual Alignment [35.33431650608965]
Cross-language pre-trained models such as multilingual BERT (mBERT) have achieved significant performance in various cross-lingual downstream NLP tasks.
This paper proposes a multi-level contrastive learning framework to further improve the cross-lingual ability of pre-trained models.
arXiv Detail & Related papers (2022-02-26T07:14:20Z) - Pair-Level Supervised Contrastive Learning for Natural Language
Inference [11.858875850000457]
We propose a Pair-level Supervised Contrastive Learning approach (PairSCL)
A contrastive learning objective is designed to distinguish the varied classes of sentence pairs by pulling those in one class together and pushing apart the pairs in other classes.
We evaluate PairSCL on two public datasets of NLI where the accuracy of PairSCL outperforms other methods by 2.1% on average.
arXiv Detail & Related papers (2022-01-26T13:34:52Z) - SLM: Learning a Discourse Language Representation with Sentence
Unshuffling [53.42814722621715]
We introduce Sentence-level Language Modeling, a new pre-training objective for learning a discourse language representation.
We show that this feature of our model improves the performance of the original BERT by large margins.
arXiv Detail & Related papers (2020-10-30T13:33:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.