Learning Fine-grained Fact-Article Correspondence in Legal Cases
- URL: http://arxiv.org/abs/2104.10726v1
- Date: Wed, 21 Apr 2021 19:06:58 GMT
- Title: Learning Fine-grained Fact-Article Correspondence in Legal Cases
- Authors: Jidong Ge, Yunyun huang, Xiaoyu Shen, Chuanyi Li, Wei Hu and Bin Luo
- Abstract summary: We create a corpus with manually annotated fact-article correspondences.
We parse articles in form of premise-conclusion pairs with random forest.
Our best system reaches an F1 score of 96.3%, making it of great potential for practical use.
- Score: 19.606628325747938
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automatically recommending relevant law articles to a given legal case has
attracted much attention as it can greatly release human labor from searching
over the large database of laws. However, current researches only support
coarse-grained recommendation where all relevant articles are predicted as a
whole without explaining which specific fact each article is relevant with.
Since one case can be formed of many supporting facts, traversing over them to
verify the correctness of recommendation results can be time-consuming. We
believe that learning fine-grained correspondence between each single fact and
law articles is crucial for an accurate and trustworthy AI system. With this
motivation, we perform a pioneering study and create a corpus with manually
annotated fact-article correspondences. We treat the learning as a text
matching task and propose a multi-level matching network to address it. To help
the model better digest the content of law articles, we parse articles in form
of premise-conclusion pairs with random forest. Experiments show that the
parsed form yielded better performance and the resulting model surpassed other
popular text matching baselines. Furthermore, we compare with previous
researches and find that establishing the fine-grained fact-article
correspondences can improve the recommendation accuracy by a large margin. Our
best system reaches an F1 score of 96.3%, making it of great potential for
practical use. It can also significantly boost the downstream
Related papers
- DELTA: Pre-train a Discriminative Encoder for Legal Case Retrieval via Structural Word Alignment [55.91429725404988]
We introduce DELTA, a discriminative model designed for legal case retrieval.
We leverage shallow decoders to create information bottlenecks, aiming to enhance the representation ability.
Our approach can outperform existing state-of-the-art methods in legal case retrieval.
arXiv Detail & Related papers (2024-03-27T10:40:14Z) - Citation Recommendation on Scholarly Legal Articles [0.0]
Citation recommendation is used within the legal domain to identify supporting arguments.
BM25 is a strong benchmark for the legal citation recommendation task.
Fine-tuning leads to considerable performance increases in pre-trained models.
arXiv Detail & Related papers (2023-11-10T07:11:55Z) - CausalCite: A Causal Formulation of Paper Citations [80.82622421055734]
CausalCite is a new way to measure the significance of a paper by assessing the causal impact of the paper on its follow-up papers.
It is based on a novel causal inference method, TextMatch, which adapts the traditional matching framework to high-dimensional text embeddings.
We demonstrate the effectiveness of CausalCite on various criteria, such as high correlation with paper impact as reported by scientific experts.
arXiv Detail & Related papers (2023-11-05T23:09:39Z) - Attentive Deep Neural Networks for Legal Document Retrieval [2.4350217735794337]
We study the use of attentive neural network-based text representation for statute law document retrieval.
We develop two hierarchical architectures with sparse attention to represent long sentences and articles, and we name them Attentive CNN and Paraformer.
Experimental results show that Attentive neural methods substantially outperform non-neural methods in terms of retrieval performance across datasets and languages.
arXiv Detail & Related papers (2022-12-13T01:37:27Z) - GERE: Generative Evidence Retrieval for Fact Verification [57.78768817972026]
We propose GERE, the first system that retrieves evidences in a generative fashion.
The experimental results on the FEVER dataset show that GERE achieves significant improvements over the state-of-the-art baselines.
arXiv Detail & Related papers (2022-04-12T03:49:35Z) - Assisting the Human Fact-Checkers: Detecting All Previously Fact-Checked
Claims in a Document [27.076320857009655]
Given an input document, it aims to detect all sentences that contain a claim that can be verified by some previously fact-checked claims.
The output is a re-ranked list of the document sentences, so that those that can be verified are ranked as high as possible.
Our analysis demonstrates the importance of modeling text similarity and stance, while also taking into account the veracity of the retrieved previously fact-checked claims.
arXiv Detail & Related papers (2021-09-14T13:46:52Z) - Toward the Understanding of Deep Text Matching Models for Information
Retrieval [72.72380690535766]
This paper aims at testing whether existing deep text matching methods satisfy some fundamental gradients in information retrieval.
Specifically, four attributions are used in our study, i.e., term frequency constraint, term discrimination constraint, length normalization constraints, and TF-length constraint.
Experimental results on LETOR 4.0 and MS Marco show that all the investigated deep text matching methods satisfy the above constraints with high probabilities in statistics.
arXiv Detail & Related papers (2021-08-16T13:33:15Z) - Evaluating Document Representations for Content-based Legal Literature
Recommendations [6.4815284696225905]
Legal recommender systems are typically evaluated in small-scale user study without any public available benchmark datasets.
We evaluate text-based (e.g., fastText, Transformers), citation-based (e.g., DeepWalk, Poincar'e), and hybrid methods.
Our experiments show that document representations from averaged fastText word vectors (trained on legal corpora) yield the best results.
arXiv Detail & Related papers (2021-04-28T15:48:19Z) - Abstractive Summarization of Spoken and Written Instructions with BERT [66.14755043607776]
We present the first application of the BERTSum model to conversational language.
We generate abstractive summaries of narrated instructional videos across a wide variety of topics.
We envision this integrated as a feature in intelligent virtual assistants, enabling them to summarize both written and spoken instructional content upon request.
arXiv Detail & Related papers (2020-08-21T20:59:34Z) - Learning Neural Textual Representations for Citation Recommendation [7.227232362460348]
We propose a novel approach to citation recommendation using a deep sequential representation of the documents (Sentence-BERT) cascaded with Siamese and triplet networks in a submodular scoring function.
To the best of our knowledge, this is the first approach to combine deep representations and submodular selection for a task of citation recommendation.
arXiv Detail & Related papers (2020-07-08T12:38:50Z) - Context-Based Quotation Recommendation [60.93257124507105]
We propose a novel context-aware quote recommendation system.
It generates a ranked list of quotable paragraphs and spans of tokens from a given source document.
We conduct experiments on a collection of speech transcripts and associated news articles.
arXiv Detail & Related papers (2020-05-17T17:49:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.