Neuro-Argumentative Learning with Case-Based Reasoning
- URL: http://arxiv.org/abs/2505.15742v1
- Date: Wed, 21 May 2025 16:49:47 GMT
- Title: Neuro-Argumentative Learning with Case-Based Reasoning
- Authors: Adam Gould, Francesca Toni,
- Abstract summary: We introduce Gradual Abstract Argumentation for Case-Based Reasoning (Gradual AA-CBR), a data-driven, neurosymbolic classification model.<n>Each argument in the debate is an observed case from the training data, favouring their labelling.<n>Cases attack or support those with opposing or agreeing labellings, with the strength of each argument and relationship learned through gradient-based methods.
- Score: 12.489784979345654
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We introduce Gradual Abstract Argumentation for Case-Based Reasoning (Gradual AA-CBR), a data-driven, neurosymbolic classification model in which the outcome is determined by an argumentation debate structure that is learned simultaneously with neural-based feature extractors. Each argument in the debate is an observed case from the training data, favouring their labelling. Cases attack or support those with opposing or agreeing labellings, with the strength of each argument and relationship learned through gradient-based methods. This argumentation debate structure provides human-aligned reasoning, improving model interpretability compared to traditional neural networks (NNs). Unlike the existing purely symbolic variant, Abstract Argumentation for Case-Based Reasoning (AA-CBR), Gradual AA-CBR is capable of multi-class classification, automatic learning of feature and data point importance, assigning uncertainty values to outcomes, using all available data points, and does not require binary features. We show that Gradual AA-CBR performs comparably to NNs whilst significantly outperforming existing AA-CBR formulations.
Related papers
- SIC: Similarity-Based Interpretable Image Classification with Neural Networks [3.0248879829045388]
We introduce SIC, a neural network that provides local and global explanations of its decision-making process.<n>We evaluate SIC on three tasks: fine-grained classification on Stanford Dogs and FunnyBirds, multi-label classification on Pascal VOC, and pathology detection on the RSNA dataset.
arXiv Detail & Related papers (2025-01-28T22:39:03Z) - Preference-Based Abstract Argumentation for Case-Based Reasoning (with Appendix) [9.5382175632919]
We introduce Preference-Based Abstract Argumentation for Case-Based Reasoning (which we call AA-CBR-P)
This allows users to define multiple approaches to compare cases with an ordering that specifies their preference over these comparison approaches.
We show empirically that our approach outperforms other interpretable machine learning models on a real-world medical dataset.
arXiv Detail & Related papers (2024-07-31T18:31:04Z) - Efficient argument classification with compact language models and ChatGPT-4 refinements [0.0]
This paper presents comparative studies between a few deep learning-based models in argument mining.
The main novelty of this paper is the ensemble model which is based on BERT architecture and ChatGPT-4 as fine tuning model.
The presented results show that BERT+ChatGPT-4 outperforms the rest of the models including other Transformer-based and LSTM-based models.
arXiv Detail & Related papers (2024-03-20T16:24:10Z) - Technical Report on the Learning of Case Relevance in Case-Based
Reasoning with Abstract Argumentation [14.755026411356315]
We show how relevance can be learnt automatically in practice with the help of decision trees.
We also show that AA-CBR with decision-tree-based learning of case relevance results in a more compact representation than their decision tree counterparts.
arXiv Detail & Related papers (2023-10-30T15:01:41Z) - SocREval: Large Language Models with the Socratic Method for Reference-Free Reasoning Evaluation [78.23119125463964]
We develop SocREval, a novel approach for prompt design in reference-free reasoning evaluation.<n>SocREval significantly improves GPT-4's performance, surpassing existing reference-free and reference-based reasoning evaluation metrics.
arXiv Detail & Related papers (2023-09-29T18:25:46Z) - Machine Reading Comprehension using Case-based Reasoning [92.51061570746077]
We present an accurate and interpretable method for answer extraction in machine reading comprehension.
Our method builds upon the hypothesis that contextualized answers to similar questions share semantic similarities with each other.
arXiv Detail & Related papers (2023-05-24T07:09:56Z) - Neural Unsupervised Semantic Role Labeling [48.69930912510414]
We present the first neural unsupervised model for semantic role labeling.
We decompose the task as two argument related subtasks, identification and clustering.
Experiments on CoNLL-2009 English dataset demonstrate that our model outperforms previous state-of-the-art baseline.
arXiv Detail & Related papers (2021-04-19T04:50:16Z) - Explaining and Improving Model Behavior with k Nearest Neighbor
Representations [107.24850861390196]
We propose using k nearest neighbor representations to identify training examples responsible for a model's predictions.
We show that kNN representations are effective at uncovering learned spurious associations.
Our results indicate that the kNN approach makes the finetuned model more robust to adversarial inputs.
arXiv Detail & Related papers (2020-10-18T16:55:25Z) - High-order Semantic Role Labeling [86.29371274587146]
This paper introduces a high-order graph structure for the neural semantic role labeling model.
It enables the model to explicitly consider not only the isolated predicate-argument pairs but also the interaction between the predicate-argument pairs.
Experimental results on 7 languages of the CoNLL-2009 benchmark show that the high-order structural learning techniques are beneficial to the strong performing SRL models.
arXiv Detail & Related papers (2020-10-09T15:33:54Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.