THUIR@COLIEE-2020: Leveraging Semantic Understanding and Exact Matching
for Legal Case Retrieval and Entailment
- URL: http://arxiv.org/abs/2012.13102v1
- Date: Thu, 24 Dec 2020 04:59:45 GMT
- Title: THUIR@COLIEE-2020: Leveraging Semantic Understanding and Exact Matching
for Legal Case Retrieval and Entailment
- Authors: Yunqiu Shao, Bulou Liu, Jiaxin Mao, Yiqun Liu, Min Zhang, Shaoping Ma
- Abstract summary: We present our methodologies for tackling the challenges of legal case retrieval and entailment.
We participated in the two case law tasks, i.e., the legal case retrieval task and the legal case entailment task.
In both tasks, we employed the neural models for semantic understanding and the traditional retrieval models for exact matching.
- Score: 41.51705651274111
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we present our methodologies for tackling the challenges of
legal case retrieval and entailment in the Competition on Legal Information
Extraction / Entailment 2020 (COLIEE-2020). We participated in the two case law
tasks, i.e., the legal case retrieval task and the legal case entailment task.
Task 1 (the retrieval task) aims to automatically identify supporting cases
from the case law corpus given a new case, and Task 2 (the entailment task) to
identify specific paragraphs that entail the decision of a new case in a
relevant case. In both tasks, we employed the neural models for semantic
understanding and the traditional retrieval models for exact matching. As a
result, our team (TLIR) ranked 2nd among all of the teams in Task 1 and 3rd
among teams in Task 2. Experimental results suggest that combing models of
semantic understanding and exact matching benefits the legal case retrieval
task while the legal case entailment task relies more on semantic
understanding.
Related papers
- LawLLM: Law Large Language Model for the US Legal System [43.13850456765944]
We introduce the Law Large Language Model (LawLLM), a multi-task model specifically designed for the US legal domain.
LawLLM excels at Similar Case Retrieval (SCR), Precedent Case Recommendation (PCR), and Legal Judgment Prediction (LJP)
We propose customized data preprocessing techniques for each task that transform raw legal data into a trainable format.
arXiv Detail & Related papers (2024-07-27T21:51:30Z) - DELTA: Pre-train a Discriminative Encoder for Legal Case Retrieval via Structural Word Alignment [55.91429725404988]
We introduce DELTA, a discriminative model designed for legal case retrieval.
We leverage shallow decoders to create information bottlenecks, aiming to enhance the representation ability.
Our approach can outperform existing state-of-the-art methods in legal case retrieval.
arXiv Detail & Related papers (2024-03-27T10:40:14Z) - MUSER: A Multi-View Similar Case Retrieval Dataset [65.36779942237357]
Similar case retrieval (SCR) is a representative legal AI application that plays a pivotal role in promoting judicial fairness.
Existing SCR datasets only focus on the fact description section when judging the similarity between cases.
We present M, a similar case retrieval dataset based on multi-view similarity measurement and comprehensive legal element with sentence-level legal element annotations.
arXiv Detail & Related papers (2023-10-24T08:17:11Z) - An Intent Taxonomy of Legal Case Retrieval [43.22489520922202]
Legal case retrieval is a special Information Retrieval(IR) task focusing on legal case documents.
We present a novel hierarchical intent taxonomy of legal case retrieval.
We reveal significant differences in user behavior and satisfaction under different search intents in legal case retrieval.
arXiv Detail & Related papers (2023-07-25T07:27:32Z) - SAILER: Structure-aware Pre-trained Language Model for Legal Case
Retrieval [75.05173891207214]
Legal case retrieval plays a core role in the intelligent legal system.
Most existing language models have difficulty understanding the long-distance dependencies between different structures.
We propose a new Structure-Aware pre-traIned language model for LEgal case Retrieval.
arXiv Detail & Related papers (2023-04-22T10:47:01Z) - SemEval 2023 Task 6: LegalEval - Understanding Legal Texts [2.172613863157655]
There is a need for developing NLP-based techniques for processing and automatically understanding legal documents.
LegalEval task has three sub-tasks: Task-A (Rhetorical Roles Labeling) is about automatically structuring legal documents into semantically coherent units, Task-B (Legal Named Entity Recognition) deals with identifying relevant entities in a legal document, Task-C (Court Judgement Prediction with Explanation) explores the possibility of automatically predicting the outcome of a legal case.
In each of the sub-tasks, the proposed systems outperformed the baselines; however, there is a lot of scope for
arXiv Detail & Related papers (2023-04-19T10:28:32Z) - Legal Element-oriented Modeling with Multi-view Contrastive Learning for
Legal Case Retrieval [3.909749182759558]
We propose an interaction-focused network for legal case retrieval with a multi-view contrastive learning objective.
Case-view contrastive learning minimizes the hidden space distance between relevant legal case representations.
We employ a legal element knowledge-aware indicator to detect legal elements of cases.
arXiv Detail & Related papers (2022-10-11T06:47:23Z) - nigam@COLIEE-22: Legal Case Retrieval and Entailment using Cascading of
Lexical and Semantic-based models [0.951828574518325]
This paper describes our submission to the Competition on Legal Information Extraction/Entailment 2022 (COLIEE-2022) workshop on case law competition for tasks 1 and 2.
We employed the neural models Sentence-BERT and Sent2Vec for semantic understanding and the traditional retrieval model BM25 for exact matching in both tasks.
arXiv Detail & Related papers (2022-04-16T18:10:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.