Pre-trained Sentence Embeddings for Implicit Discourse Relation
Classification
- URL: http://arxiv.org/abs/2210.11005v1
- Date: Thu, 20 Oct 2022 04:17:03 GMT
- Title: Pre-trained Sentence Embeddings for Implicit Discourse Relation
Classification
- Authors: Murali Raghu Babu Balusu, Yangfeng Ji and Jacob Eisenstein
- Abstract summary: Implicit discourse relations bind smaller linguistic units into coherent texts.
We explore the utility of pre-trained sentence embeddings as base representations in a neural network for implicit discourse relation sense classification.
- Score: 26.973476248983477
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Implicit discourse relations bind smaller linguistic units into coherent
texts. Automatic sense prediction for implicit relations is hard, because it
requires understanding the semantics of the linked arguments. Furthermore,
annotated datasets contain relatively few labeled examples, due to the scale of
the phenomenon: on average each discourse relation encompasses several dozen
words. In this paper, we explore the utility of pre-trained sentence embeddings
as base representations in a neural network for implicit discourse relation
sense classification. We present a series of experiments using both supervised
end-to-end trained models and pre-trained sentence encoding techniques -
SkipThought, Sent2vec and Infersent. The pre-trained embeddings are competitive
with the end-to-end model, and the approaches are complementary, with combined
models yielding significant performance improvements on two of the three
evaluations.
Related papers
- Manual Verbalizer Enrichment for Few-Shot Text Classification [1.860409237919611]
acrshortmave is an approach for verbalizer construction by enrichment of class labels.
Our model achieves state-of-the-art results while using significantly fewer resources.
arXiv Detail & Related papers (2024-10-08T16:16:47Z) - How Well Do Text Embedding Models Understand Syntax? [50.440590035493074]
The ability of text embedding models to generalize across a wide range of syntactic contexts remains under-explored.
Our findings reveal that existing text embedding models have not sufficiently addressed these syntactic understanding challenges.
We propose strategies to augment the generalization ability of text embedding models in diverse syntactic scenarios.
arXiv Detail & Related papers (2023-11-14T08:51:00Z) - Annotation-Inspired Implicit Discourse Relation Classification with
Auxiliary Discourse Connective Generation [14.792252724959383]
Implicit discourse relation classification is a challenging task due to the absence of discourse connectives.
We design an end-to-end neural model to explicitly generate discourse connectives for the task, inspired by the annotation process of PDTB.
Specifically, our model jointly learns to generate discourse connectives between arguments and predict discourse relations based on the arguments and the generated connectives.
arXiv Detail & Related papers (2023-06-10T16:38:46Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - Distant finetuning with discourse relations for stance classification [55.131676584455306]
We propose a new method to extract data with silver labels from raw text to finetune a model for stance classification.
We also propose a 3-stage training framework where the noisy level in the data used for finetuning decreases over different stages.
Our approach ranks 1st among 26 competing teams in the stance classification track of the NLPCC 2021 shared task Argumentative Text Understanding for AI Debater.
arXiv Detail & Related papers (2022-04-27T04:24:35Z) - Analysis of Joint Speech-Text Embeddings for Semantic Matching [3.6423306784901235]
We study a joint speech-text embedding space trained for semantic matching by minimizing the distance between paired utterance and transcription inputs.
We extend our method to incorporate automatic speech recognition through both pretraining and multitask scenarios.
arXiv Detail & Related papers (2022-04-04T04:50:32Z) - Avoiding Inference Heuristics in Few-shot Prompt-based Finetuning [57.4036085386653]
We show that prompt-based models for sentence pair classification tasks still suffer from a common pitfall of adopting inferences based on lexical overlap.
We then show that adding a regularization that preserves pretraining weights is effective in mitigating this destructive tendency of few-shot finetuning.
arXiv Detail & Related papers (2021-09-09T10:10:29Z) - Let's be explicit about that: Distant supervision for implicit discourse
relation classification via connective prediction [0.0]
In implicit discourse relation classification, we want to predict the relation between adjacent sentences in the absence of any overt discourse connectives.
We sidestep the lack of data through explicitation of implicit relations to reduce the task to two sub-problems: language modeling and explicit discourse relation classification.
Our experimental results show that this method can even marginally outperform the state-of-the-art, in spite of being much simpler than alternative models of comparable performance.
arXiv Detail & Related papers (2021-06-06T17:57:32Z) - Syntax-Enhanced Pre-trained Model [49.1659635460369]
We study the problem of leveraging the syntactic structure of text to enhance pre-trained models such as BERT and RoBERTa.
Existing methods utilize syntax of text either in the pre-training stage or in the fine-tuning stage, so that they suffer from discrepancy between the two stages.
We present a model that utilizes the syntax of text in both pre-training and fine-tuning stages.
arXiv Detail & Related papers (2020-12-28T06:48:04Z) - Weakly-Supervised Aspect-Based Sentiment Analysis via Joint
Aspect-Sentiment Topic Embedding [71.2260967797055]
We propose a weakly-supervised approach for aspect-based sentiment analysis.
We learn sentiment, aspect> joint topic embeddings in the word embedding space.
We then use neural models to generalize the word-level discriminative information.
arXiv Detail & Related papers (2020-10-13T21:33:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.