It Takes Two Flints to Make a Fire: Multitask Learning of Neural
Relation and Explanation Classifiers
- URL: http://arxiv.org/abs/2204.11424v1
- Date: Mon, 25 Apr 2022 03:53:12 GMT
- Title: It Takes Two Flints to Make a Fire: Multitask Learning of Neural
Relation and Explanation Classifiers
- Authors: Zheng Tang, Mihai Surdeanu
- Abstract summary: We propose an explainable approach for relation extraction that mitigates the tension between generalization and explainability.
Our approach uses a multi-task learning architecture, which jointly trains a classifier for relation extraction.
We convert the model outputs to rules to bring global explanations to this approach.
- Score: 40.666590079580544
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose an explainable approach for relation extraction that mitigates the
tension between generalization and explainability by jointly training for the
two goals. Our approach uses a multi-task learning architecture, which jointly
trains a classifier for relation extraction, and a sequence model that labels
words in the context of the relation that explain the decisions of the relation
classifier. We also convert the model outputs to rules to bring global
explanations to this approach. This sequence model is trained using a hybrid
strategy: supervised, when supervision from pre-existing patterns is available,
and semi-supervised otherwise. In the latter situation, we treat the sequence
model's labels as latent variables, and learn the best assignment that
maximizes the performance of the relation classifier. We evaluate the proposed
approach on the two datasets and show that the sequence model provides labels
that serve as accurate explanations for the relation classifier's decisions,
and, importantly, that the joint training generally improves the performance of
the relation classifier. We also evaluate the performance of the generated
rules and show that the new rules are great add-on to the manual rules and
bring the rule-based system much closer to the neural models.
Related papers
- Best of Both Worlds: A Pliable and Generalizable Neuro-Symbolic Approach
for Relation Classification [17.398872494876365]
This paper introduces a novel neuro-symbolic architecture for relation classification (RC)
It combines rule-based methods with contemporary deep learning techniques.
We show that our proposed method outperforms previous state-of-the-art models in three out of four settings.
arXiv Detail & Related papers (2024-03-05T20:08:32Z) - Graph-based Time Series Clustering for End-to-End Hierarchical Forecasting [18.069747511100132]
Relationships among time series can be exploited as inductive biases in learning effective forecasting models.
We propose a graph-based methodology to unify relational and hierarchical inductive biases.
arXiv Detail & Related papers (2023-05-30T16:27:25Z) - Mitigating Catastrophic Forgetting in Task-Incremental Continual
Learning with Adaptive Classification Criterion [50.03041373044267]
We propose a Supervised Contrastive learning framework with adaptive classification criterion for Continual Learning.
Experiments show that CFL achieves state-of-the-art performance and has a stronger ability to overcome compared with the classification baselines.
arXiv Detail & Related papers (2023-05-20T19:22:40Z) - HiURE: Hierarchical Exemplar Contrastive Learning for Unsupervised
Relation Extraction [60.80849503639896]
Unsupervised relation extraction aims to extract the relationship between entities from natural language sentences without prior information on relational scope or distribution.
We propose a novel contrastive learning framework named HiURE, which has the capability to derive hierarchical signals from relational feature space using cross hierarchy attention.
Experimental results on two public datasets demonstrate the advanced effectiveness and robustness of HiURE on unsupervised relation extraction when compared with state-of-the-art models.
arXiv Detail & Related papers (2022-05-04T17:56:48Z) - Generative Relation Linking for Question Answering over Knowledge Bases [12.778133758613773]
We propose a novel approach for relation linking framing it as a generative problem.
We extend such sequence-to-sequence models with the idea of infusing structured data from the target knowledge base.
We train the model with the aim to generate a structured output consisting of a list of argument-relation pairs, enabling a knowledge validation step.
arXiv Detail & Related papers (2021-08-16T20:33:43Z) - Refining Neural Networks with Compositional Explanations [31.84868477264624]
We propose to refine a learned model by collecting human-provided compositional explanations on the models' failure cases.
We demonstrate the effectiveness of the proposed approach on two text classification tasks.
arXiv Detail & Related papers (2021-03-18T17:48:54Z) - Clustering-based Unsupervised Generative Relation Extraction [3.342376225738321]
We propose a Clustering-based Unsupervised generative Relation Extraction framework (CURE)
We use an "Encoder-Decoder" architecture to perform self-supervised learning so the encoder can extract relation information.
Our model performs better than state-of-the-art models on both New York Times (NYT) and United Nations Parallel Corpus (UNPC) standard datasets.
arXiv Detail & Related papers (2020-09-26T20:36:40Z) - S^3-Rec: Self-Supervised Learning for Sequential Recommendation with
Mutual Information Maximization [104.87483578308526]
We propose the model S3-Rec, which stands for Self-Supervised learning for Sequential Recommendation.
For our task, we devise four auxiliary self-supervised objectives to learn the correlations among attribute, item, subsequence, and sequence.
Extensive experiments conducted on six real-world datasets demonstrate the superiority of our proposed method over existing state-of-the-art methods.
arXiv Detail & Related papers (2020-08-18T11:44:10Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z) - Adaptive Correlated Monte Carlo for Contextual Categorical Sequence
Generation [77.7420231319632]
We adapt contextual generation of categorical sequences to a policy gradient estimator, which evaluates a set of correlated Monte Carlo (MC) rollouts for variance control.
We also demonstrate the use of correlated MC rollouts for binary-tree softmax models, which reduce the high generation cost in large vocabulary scenarios.
arXiv Detail & Related papers (2019-12-31T03:01:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.