On the Importance of Word and Sentence Representation Learning in
Implicit Discourse Relation Classification
- URL: http://arxiv.org/abs/2004.12617v2
- Date: Tue, 28 Apr 2020 15:49:48 GMT
- Title: On the Importance of Word and Sentence Representation Learning in
Implicit Discourse Relation Classification
- Authors: Xin Liu, Jiefu Ou, Yangqiu Song, Xin Jiang
- Abstract summary: Implicit discourse relation classification is one of the most difficult parts in shallow discourse parsing.
We argue that a powerful contextualized representation module, a bilateral multi-perspective matching module, and a global information fusion module are all important to implicit discourse analysis.
- Score: 43.483855615908695
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Implicit discourse relation classification is one of the most difficult parts
in shallow discourse parsing as the relation prediction without explicit
connectives requires the language understanding at both the text span level and
the sentence level. Previous studies mainly focus on the interactions between
two arguments. We argue that a powerful contextualized representation module, a
bilateral multi-perspective matching module, and a global information fusion
module are all important to implicit discourse analysis. We propose a novel
model to combine these modules together. Extensive experiments show that our
proposed model outperforms BERT and other state-of-the-art systems on the PDTB
dataset by around 8% and CoNLL 2016 datasets around 16%. We also analyze the
effectiveness of different modules in the implicit discourse relation
classification task and demonstrate how different levels of representation
learning can affect the results.
Related papers
- Enhancing Argument Structure Extraction with Efficient Leverage of
Contextual Information [79.06082391992545]
We propose an Efficient Context-aware model (ECASE) that fully exploits contextual information.
We introduce a sequence-attention module and distance-weighted similarity loss to aggregate contextual information and argumentative information.
Our experiments on five datasets from various domains demonstrate that our model achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-10-08T08:47:10Z) - Multi-Grained Multimodal Interaction Network for Entity Linking [65.30260033700338]
Multimodal entity linking task aims at resolving ambiguous mentions to a multimodal knowledge graph.
We propose a novel Multi-GraIned Multimodal InteraCtion Network $textbf(MIMIC)$ framework for solving the MEL task.
arXiv Detail & Related papers (2023-07-19T02:11:19Z) - Disco-Bench: A Discourse-Aware Evaluation Benchmark for Language
Modelling [70.23876429382969]
We propose a benchmark that can evaluate intra-sentence discourse properties across a diverse set of NLP tasks.
Disco-Bench consists of 9 document-level testsets in the literature domain, which contain rich discourse phenomena.
For linguistic analysis, we also design a diagnostic test suite that can examine whether the target models learn discourse knowledge.
arXiv Detail & Related papers (2023-07-16T15:18:25Z) - Topics in the Haystack: Extracting and Evaluating Topics beyond
Coherence [0.0]
We propose a method that incorporates a deeper understanding of both sentence and document themes.
This allows our model to detect latent topics that may include uncommon words or neologisms.
We present correlation coefficients with human identification of intruder words and achieve near-human level results at the word-intrusion task.
arXiv Detail & Related papers (2023-03-30T12:24:25Z) - Model and Evaluation: Towards Fairness in Multilingual Text
Classification [6.697876965452054]
We propose a debiasing framework for multilingual text classification based on contrastive learning.
The model contains four modules: multilingual text representation module, language fusion module, text debiasing module, and text classification module.
We propose a multi-dimensional fairness evaluation framework for multilingual text classification, which evaluates the model's monolingual equality difference, multilingual equality difference, multilingual equality performance difference, and destructiveness of the fairness strategy.
arXiv Detail & Related papers (2023-03-28T03:00:01Z) - Improve Discourse Dependency Parsing with Contextualized Representations [28.916249926065273]
We propose to take advantage of transformers to encode contextualized representations of units of different levels.
Motivated by the observation of writing patterns commonly shared across articles, we propose a novel method that treats discourse relation identification as a sequence labelling task.
arXiv Detail & Related papers (2022-05-04T14:35:38Z) - Incorporating Linguistic Knowledge for Abstractive Multi-document
Summarization [20.572283625521784]
We develop a neural network based abstractive multi-document summarization (MDS) model.
We process the dependency information into the linguistic-guided attention mechanism.
With the help of linguistic signals, sentence-level relations can be correctly captured.
arXiv Detail & Related papers (2021-09-23T08:13:35Z) - Prototypical Representation Learning for Relation Extraction [56.501332067073065]
This paper aims to learn predictive, interpretable, and robust relation representations from distantly-labeled data.
We learn prototypes for each relation from contextual information to best explore the intrinsic semantics of relations.
Results on several relation learning tasks show that our model significantly outperforms the previous state-of-the-art relational models.
arXiv Detail & Related papers (2021-03-22T08:11:43Z) - SPLAT: Speech-Language Joint Pre-Training for Spoken Language
Understanding [61.02342238771685]
Spoken language understanding requires a model to analyze input acoustic signal to understand its linguistic content and make predictions.
Various pre-training methods have been proposed to learn rich representations from large-scale unannotated speech and text.
We propose a novel semi-supervised learning framework, SPLAT, to jointly pre-train the speech and language modules.
arXiv Detail & Related papers (2020-10-05T19:29:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.