Multi-granularity Argument Mining in Legal Texts
- URL: http://arxiv.org/abs/2210.09472v2
- Date: Wed, 19 Oct 2022 15:00:55 GMT
- Title: Multi-granularity Argument Mining in Legal Texts
- Authors: Huihui Xu and Kevin Ashley
- Abstract summary: We conceptualize argument mining as a token-level (i.e., word-level) classification problem.
Results show that token-level text classification identifies certain legal argument elements more accurately than sentence-level text classification.
- Score: 0.913755431537592
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper, we explore legal argument mining using multiple levels of
granularity. Argument mining has usually been conceptualized as a sentence
classification problem. In this work, we conceptualize argument mining as a
token-level (i.e., word-level) classification problem. We use a Longformer
model to classify the tokens. Results show that token-level text classification
identifies certain legal argument elements more accurately than sentence-level
text classification. Token-level classification also provides greater
flexibility to analyze legal texts and to gain more insight into what the model
focuses on when processing a large amount of input data.
Related papers
- Language Models for Text Classification: Is In-Context Learning Enough? [54.869097980761595]
Recent foundational language models have shown state-of-the-art performance in many NLP tasks in zero- and few-shot settings.
An advantage of these models over more standard approaches is the ability to understand instructions written in natural language (prompts)
This makes them suitable for addressing text classification problems for domains with limited amounts of annotated instances.
arXiv Detail & Related papers (2024-03-26T12:47:39Z) - Argue with Me Tersely: Towards Sentence-Level Counter-Argument
Generation [62.069374456021016]
We present the ArgTersely benchmark for sentence-level counter-argument generation.
We also propose Arg-LlaMA for generating high-quality counter-argument.
arXiv Detail & Related papers (2023-12-21T06:51:34Z) - Like a Good Nearest Neighbor: Practical Content Moderation and Text
Classification [66.02091763340094]
Like a Good Nearest Neighbor (LaGoNN) is a modification to SetFit that introduces no learnable parameters but alters input text with information from its nearest neighbor.
LaGoNN is effective at flagging undesirable content and text classification, and improves the performance of SetFit.
arXiv Detail & Related papers (2023-02-17T15:43:29Z) - Zero-Shot Classification by Logical Reasoning on Natural Language
Explanations [56.42922904777717]
We propose the framework CLORE (Classification by LOgical Reasoning on Explanations)
CLORE parses explanations into logical structures and then explicitly reasons along thess structures on the input to produce a classification score.
We also demonstrate that our framework can be extended to zero-shot classification on visual modality.
arXiv Detail & Related papers (2022-11-07T01:05:11Z) - LIME: Weakly-Supervised Text Classification Without Seeds [1.2691047660244335]
In weakly-supervised text classification, only label names act as sources of supervision.
We present LIME, a framework for weakly-supervised text classification.
We find that combining weakly-supervised classification and textual entailment mitigates shortcomings of both.
arXiv Detail & Related papers (2022-10-13T04:28:28Z) - Perturbations and Subpopulations for Testing Robustness in Token-Based
Argument Unit Recognition [6.502694770864571]
Argument Unit Recognition and Classification aims at identifying argument units from text and classifying them as pro or against.
One of the design choices that need to be made when developing systems for this task is what the unit of classification should be: segments of tokens or full sentences.
Previous research suggests that fine-tuning language models on the token-level yields more robust results for classifying sentences compared to training on sentences directly.
We reproduce the study that originally made this claim and further investigate what exactly token-based systems learned better compared to sentence-based ones.
arXiv Detail & Related papers (2022-09-29T13:44:28Z) - More Than Words: Towards Better Quality Interpretations of Text
Classifiers [16.66535643383862]
We show that token-based interpretability, while being a convenient first choice given the input interfaces of the ML models, is not the most effective one in all situations.
We show that higher-level feature attributions offer several advantages: 1) they are more robust as measured by the randomization tests, 2) they lead to lower variability when using approximation-based methods like SHAP, and 3) they are more intelligible to humans in situations where the linguistic coherence resides at a higher level.
arXiv Detail & Related papers (2021-12-23T10:18:50Z) - Clustering and Network Analysis for the Embedding Spaces of Sentences
and Sub-Sentences [69.3939291118954]
This paper reports research on a set of comprehensive clustering and network analyses targeting sentence and sub-sentence embedding spaces.
Results show that one method generates the most clusterable embeddings.
In general, the embeddings of span sub-sentences have better clustering properties than the original sentences.
arXiv Detail & Related papers (2021-10-02T00:47:35Z) - "Sharks are not the threat humans are": Argument Component Segmentation
in School Student Essays [3.632177840361928]
We apply a token-level classification to identify claim and premise tokens from a new corpus of argumentative essays written by middle school students.
We demonstrate that a BERT-based multi-task learning architecture (i.e., token and sentence level classification) adaptively pretrained on a relevant unlabeled dataset obtains the best results.
arXiv Detail & Related papers (2021-03-08T02:40:07Z) - Narrative Incoherence Detection [76.43894977558811]
We propose the task of narrative incoherence detection as a new arena for inter-sentential semantic understanding.
Given a multi-sentence narrative, decide whether there exist any semantic discrepancies in the narrative flow.
arXiv Detail & Related papers (2020-12-21T07:18:08Z) - Rank over Class: The Untapped Potential of Ranking in Natural Language
Processing [8.637110868126546]
We argue that many tasks which are currently addressed using classification are in fact being shoehorned into a classification mould.
We propose a novel end-to-end ranking approach consisting of a Transformer network responsible for producing representations for a pair of text sequences.
In an experiment on a heavily-skewed sentiment analysis dataset, converting ranking results to classification labels yields an approximately 22% improvement over state-of-the-art text classification.
arXiv Detail & Related papers (2020-09-10T22:18:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.