Best of Both Worlds: A Pliable and Generalizable Neuro-Symbolic Approach
for Relation Classification
- URL: http://arxiv.org/abs/2403.03305v1
- Date: Tue, 5 Mar 2024 20:08:32 GMT
- Title: Best of Both Worlds: A Pliable and Generalizable Neuro-Symbolic Approach
for Relation Classification
- Authors: Robert Vacareanu, Fahmida Alam, Md Asiful Islam, Haris Riaz, Mihai
Surdeanu
- Abstract summary: This paper introduces a novel neuro-symbolic architecture for relation classification (RC)
It combines rule-based methods with contemporary deep learning techniques.
We show that our proposed method outperforms previous state-of-the-art models in three out of four settings.
- Score: 17.398872494876365
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper introduces a novel neuro-symbolic architecture for relation
classification (RC) that combines rule-based methods with contemporary deep
learning techniques. This approach capitalizes on the strengths of both
paradigms: the adaptability of rule-based systems and the generalization power
of neural networks. Our architecture consists of two components: a declarative
rule-based model for transparent classification and a neural component to
enhance rule generalizability through semantic text matching. Notably, our
semantic matcher is trained in an unsupervised domain-agnostic way, solely with
synthetic data. Further, these components are loosely coupled, allowing for
rule modifications without retraining the semantic matcher. In our evaluation,
we focused on two few-shot relation classification datasets: Few-Shot TACRED
and a Few-Shot version of NYT29. We show that our proposed method outperforms
previous state-of-the-art models in three out of four settings, despite not
seeing any human-annotated training data. Further, we show that our approach
remains modular and pliable, i.e., the corresponding rules can be locally
modified to improve the overall model. Human interventions to the rules for the
TACRED relation \texttt{org:parents} boost the performance on that relation by
as much as 26\% relative improvement, without negatively impacting the other
relations, and without retraining the semantic matching component.
Related papers
- Hierarchical Insights: Exploiting Structural Similarities for Reliable 3D Semantic Segmentation [4.480310276450028]
We propose a training strategy for a 3D LiDAR semantic segmentation model that learns structural relationships between classes through abstraction.
This is achieved by implicitly modeling these relationships using a learning rule for hierarchical multi-label classification (HMC)
Our detailed analysis demonstrates that this training strategy not only improves the model's confidence calibration but also retains additional information useful for downstream tasks such as fusion, prediction, and planning.
arXiv Detail & Related papers (2024-04-09T08:49:01Z) - Scalable Learning of Latent Language Structure With Logical Offline
Cycle Consistency [71.42261918225773]
Conceptually, LOCCO can be viewed as a form of self-learning where the semantic being trained is used to generate annotations for unlabeled text.
As an added bonus, the annotations produced by LOCCO can be trivially repurposed to train a neural text generation model.
arXiv Detail & Related papers (2023-05-31T16:47:20Z) - TWINS: A Fine-Tuning Framework for Improved Transferability of
Adversarial Robustness and Generalization [89.54947228958494]
This paper focuses on the fine-tuning of an adversarially pre-trained model in various classification tasks.
We propose a novel statistics-based approach, Two-WIng NormliSation (TWINS) fine-tuning framework.
TWINS is shown to be effective on a wide range of image classification datasets in terms of both generalization and robustness.
arXiv Detail & Related papers (2023-03-20T14:12:55Z) - Neural-based classification rule learning for sequential data [0.0]
We propose a novel differentiable fully interpretable method to discover both local and global patterns for rule-based binary classification.
It consists of a convolutional binary neural network with an interpretable neural filter and a training strategy based on dynamically-enforced sparsity.
We demonstrate the validity and usefulness of the approach on synthetic datasets and on an open-source peptides dataset.
arXiv Detail & Related papers (2023-02-22T11:05:05Z) - Variational Cross-Graph Reasoning and Adaptive Structured Semantics
Learning for Compositional Temporal Grounding [143.5927158318524]
Temporal grounding is the task of locating a specific segment from an untrimmed video according to a query sentence.
We introduce a new Compositional Temporal Grounding task and construct two new dataset splits.
We argue that the inherent structured semantics inside the videos and language is the crucial factor to achieve compositional generalization.
arXiv Detail & Related papers (2023-01-22T08:02:23Z) - Federated Learning Aggregation: New Robust Algorithms with Guarantees [63.96013144017572]
Federated learning has been recently proposed for distributed model training at the edge.
This paper presents a complete general mathematical convergence analysis to evaluate aggregation strategies in a federated learning framework.
We derive novel aggregation algorithms which are able to modify their model architecture by differentiating client contributions according to the value of their losses.
arXiv Detail & Related papers (2022-05-22T16:37:53Z) - It Takes Two Flints to Make a Fire: Multitask Learning of Neural
Relation and Explanation Classifiers [40.666590079580544]
We propose an explainable approach for relation extraction that mitigates the tension between generalization and explainability.
Our approach uses a multi-task learning architecture, which jointly trains a classifier for relation extraction.
We convert the model outputs to rules to bring global explanations to this approach.
arXiv Detail & Related papers (2022-04-25T03:53:12Z) - Integrating Regular Expressions with Neural Networks via DFA [40.09868407372605]
It is very important to integrate the rule knowledge into neural networks to build a hybrid model that achieves better performance.
Specifically, the human-designed rules are formulated as Regular Expressions (REs)
We propose to use the MDFA as an intermediate model to capture the matched RE patterns as rule-based features for each input sentence.
arXiv Detail & Related papers (2021-09-07T05:48:51Z) - Semantic Correspondence with Transformers [68.37049687360705]
We propose Cost Aggregation with Transformers (CATs) to find dense correspondences between semantically similar images.
We include appearance affinity modelling to disambiguate the initial correlation maps and multi-level aggregation.
We conduct experiments to demonstrate the effectiveness of the proposed model over the latest methods and provide extensive ablation studies.
arXiv Detail & Related papers (2021-06-04T14:39:03Z) - Bidirectional Mapping Coupled GAN for Generalized Zero-Shot Learning [7.22073260315824]
Bidirectional mapping-based generalized zero-shot learning (GZSL) methods rely on the quality of synthesized features to recognize seen and unseen data.
We learn a joint distribution of seen-unseen domains and preserving domain distinction is crucial for these methods.
In this work, we utilize the available unseen class semantics alongside seen class semantics and learn joint distribution through a strong visual-semantic coupling.
arXiv Detail & Related papers (2020-12-30T06:11:29Z) - Revisiting LSTM Networks for Semi-Supervised Text Classification via
Mixed Objective Function [106.69643619725652]
We develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results.
We report state-of-the-art results for text classification task on several benchmark datasets.
arXiv Detail & Related papers (2020-09-08T21:55:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.