Enhancing Legal Argument Mining with Domain Pre-training and Neural
Networks
- URL: http://arxiv.org/abs/2202.13457v1
- Date: Sun, 27 Feb 2022 21:24:53 GMT
- Title: Enhancing Legal Argument Mining with Domain Pre-training and Neural
Networks
- Authors: Gechuan Zhang, Paul Nulty, David Lillis
- Abstract summary: The contextual word embedding model, BERT, has proved its ability on downstream tasks with limited quantities of annotated data.
BERT and its variants help to reduce the burden of complex annotation work in many interdisciplinary research areas.
- Score: 0.45119235878273
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The contextual word embedding model, BERT, has proved its ability on
downstream tasks with limited quantities of annotated data. BERT and its
variants help to reduce the burden of complex annotation work in many
interdisciplinary research areas, for example, legal argument mining in digital
humanities. Argument mining aims to develop text analysis tools that can
automatically retrieve arguments and identify relationships between
argumentation clauses. Since argumentation is one of the key aspects of case
law, argument mining tools for legal texts are applicable to both academic and
non-academic legal research. Domain-specific BERT variants (pre-trained with
corpora from a particular background) have also achieved strong performance in
many tasks. To our knowledge, previous machine learning studies of argument
mining on judicial case law still heavily rely on statistical models. In this
paper, we provide a broad study of both classic and contextual embedding models
and their performance on practical case law from the European Court of Human
Rights (ECHR). During our study, we also explore a number of neural networks
when being combined with different embeddings. Our experiments provide a
comprehensive overview of a variety of approaches to the legal argument mining
task. We conclude that domain pre-trained transformer models have great
potential in this area, although traditional embeddings can also achieve strong
performance when combined with additional neural network layers.
Related papers
- Prompting Encoder Models for Zero-Shot Classification: A Cross-Domain Study in Italian [75.94354349994576]
This paper explores the feasibility of employing smaller, domain-specific encoder LMs alongside prompting techniques to enhance performance in specialized contexts.
Our study concentrates on the Italian bureaucratic and legal language, experimenting with both general-purpose and further pre-trained encoder-only models.
The results indicate that while further pre-trained models may show diminished robustness in general knowledge, they exhibit superior adaptability for domain-specific tasks, even in a zero-shot setting.
arXiv Detail & Related papers (2024-07-30T08:50:16Z) - DELTA: Pre-train a Discriminative Encoder for Legal Case Retrieval via Structural Word Alignment [55.91429725404988]
We introduce DELTA, a discriminative model designed for legal case retrieval.
We leverage shallow decoders to create information bottlenecks, aiming to enhance the representation ability.
Our approach can outperform existing state-of-the-art methods in legal case retrieval.
arXiv Detail & Related papers (2024-03-27T10:40:14Z) - Exploring the Potential of Large Language Models in Computational Argumentation [54.85665903448207]
Large language models (LLMs) have demonstrated impressive capabilities in understanding context and generating natural language.
This work aims to embark on an assessment of LLMs, such as ChatGPT, Flan models, and LLaMA2 models, in both zero-shot and few-shot settings.
arXiv Detail & Related papers (2023-11-15T15:12:15Z) - Enhancing Pre-Trained Language Models with Sentence Position Embeddings
for Rhetorical Roles Recognition in Legal Opinions [0.16385815610837165]
The size of legal opinions continues to grow, making it increasingly challenging to develop a model that can accurately predict the rhetorical roles of legal opinions.
We propose a novel model architecture for automatically predicting rhetorical roles using pre-trained language models (PLMs) enhanced with knowledge of sentence position information.
Based on an annotated corpus from the LegalEval@SemEval2023 competition, we demonstrate that our approach requires fewer parameters, resulting in lower computational costs.
arXiv Detail & Related papers (2023-10-08T20:33:55Z) - SAILER: Structure-aware Pre-trained Language Model for Legal Case
Retrieval [75.05173891207214]
Legal case retrieval plays a core role in the intelligent legal system.
Most existing language models have difficulty understanding the long-distance dependencies between different structures.
We propose a new Structure-Aware pre-traIned language model for LEgal case Retrieval.
arXiv Detail & Related papers (2023-04-22T10:47:01Z) - The Legal Argument Reasoning Task in Civil Procedure [2.079168053329397]
We present a new NLP task and dataset from the domain of the U.S. civil procedure.
Each instance of the dataset consists of a general introduction to the case, a particular question, and a possible solution argument.
arXiv Detail & Related papers (2022-11-05T17:41:00Z) - Revisiting the Roles of "Text" in Text Games [102.22750109468652]
This paper investigates the roles of text in the face of different reinforcement learning challenges.
We propose a simple scheme to extract relevant contextual information into an approximate state hash.
Such a lightweight plug-in achieves competitive performance with state-of-the-art text agents.
arXiv Detail & Related papers (2022-10-15T21:52:39Z) - Mining Legal Arguments in Court Decisions [43.09204050756282]
We develop a new annotation scheme for legal arguments in proceedings of the European Court of Human Rights.
Second, we compile and annotate a large corpus of 373 court decisions.
Third, we train an argument mining model that outperforms state-of-the-art models in the legal NLP domain.
arXiv Detail & Related papers (2022-08-12T08:59:55Z) - Can Unsupervised Knowledge Transfer from Social Discussions Help
Argument Mining? [25.43442712037725]
We propose a novel transfer learning strategy to overcome the challenges of unsupervised, argumentative discourse-aware knowledge.
We utilize argumentation-rich social discussions from the ChangeMyView subreddit as a source of unsupervised, argumentative discourse-aware knowledge.
We introduce a novel prompt-based strategy for inter-component relation prediction that compliments our proposed finetuning method.
arXiv Detail & Related papers (2022-03-24T06:48:56Z) - The Unreasonable Effectiveness of the Baseline: Discussing SVMs in Legal
Text Classification [0.0]
We show that a more traditional approach based on Support Vector Machine classifiers reaches competitive performance with deep learning models.
We also highlight that error reduction obtained by using specialised BERT-based models over baselines is noticeably smaller in the legal domain when compared to general language tasks.
arXiv Detail & Related papers (2021-09-15T12:05:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.