Context-Aware Legal Citation Recommendation using Deep Learning
- URL: http://arxiv.org/abs/2106.10776v1
- Date: Sun, 20 Jun 2021 23:23:11 GMT
- Title: Context-Aware Legal Citation Recommendation using Deep Learning
- Authors: Zihan Huang, Charles Low, Mengqiu Teng, Hongyi Zhang, Daniel E. Ho,
Mark S. Krass, Matthias Grabmair
- Abstract summary: Lawyers and judges spend a large amount of time researching the proper legal authority to cite while drafting decisions.
We develop a citation recommendation tool that can help improve efficiency in the process of opinion drafting.
- Score: 4.157772749568094
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Lawyers and judges spend a large amount of time researching the proper legal
authority to cite while drafting decisions. In this paper, we develop a
citation recommendation tool that can help improve efficiency in the process of
opinion drafting. We train four types of machine learning models, including a
citation-list based method (collaborative filtering) and three context-based
methods (text similarity, BiLSTM and RoBERTa classifiers). Our experiments show
that leveraging local textual context improves recommendation, and that deep
neural models achieve decent performance. We show that non-deep text-based
methods benefit from access to structured case metadata, but deep models only
benefit from such access when predicting from context of insufficient length.
We also find that, even after extensive training, RoBERTa does not outperform a
recurrent neural model, despite its benefits of pretraining. Our behavior
analysis of the RoBERTa model further shows that predictive performance is
stable across time and citation classes.
Related papers
- RAVEN: In-Context Learning with Retrieval-Augmented Encoder-Decoder Language Models [57.12888828853409]
RAVEN is a model that combines retrieval-augmented masked language modeling and prefix language modeling.
Fusion-in-Context Learning enables the model to leverage more in-context examples without requiring additional training.
Our work underscores the potential of retrieval-augmented encoder-decoder language models for in-context learning.
arXiv Detail & Related papers (2023-08-15T17:59:18Z) - Fairness-guided Few-shot Prompting for Large Language Models [93.05624064699965]
In-context learning can suffer from high instability due to variations in training examples, example order, and prompt formats.
We introduce a metric to evaluate the predictive bias of a fixed prompt against labels or a given attributes.
We propose a novel search strategy based on the greedy search to identify the near-optimal prompt for improving the performance of in-context learning.
arXiv Detail & Related papers (2023-03-23T12:28:25Z) - Efficient and Flexible Topic Modeling using Pretrained Embeddings and
Bag of Sentences [1.8592384822257952]
We propose a novel topic modeling and inference algorithm.
We leverage pre-trained sentence embeddings by combining generative process models and clustering.
TheTailor evaluation shows that our method yields state-of-the art results with relatively little computational demands.
arXiv Detail & Related papers (2023-02-06T20:13:11Z) - SLCNN: Sentence-Level Convolutional Neural Network for Text
Classification [0.0]
Convolutional neural network (CNN) has shown remarkable success in the task of text classification.
New baseline models have been studied for text classification using CNN.
Results have shown that the proposed models have better performance, particularly in the longer documents.
arXiv Detail & Related papers (2023-01-27T13:16:02Z) - HanoiT: Enhancing Context-aware Translation via Selective Context [95.93730812799798]
Context-aware neural machine translation aims to use the document-level context to improve translation quality.
The irrelevant or trivial words may bring some noise and distract the model from learning the relationship between the current sentence and the auxiliary context.
We propose a novel end-to-end encoder-decoder model with a layer-wise selection mechanism to sift and refine the long document context.
arXiv Detail & Related papers (2023-01-17T12:07:13Z) - Efficient Nearest Neighbor Language Models [114.40866461741795]
Non-parametric neural language models (NLMs) learn predictive distributions of text utilizing an external datastore.
We show how to achieve up to a 6x speed-up in inference speed while retaining comparable performance.
arXiv Detail & Related papers (2021-09-09T12:32:28Z) - Evaluating Document Representations for Content-based Legal Literature
Recommendations [6.4815284696225905]
Legal recommender systems are typically evaluated in small-scale user study without any public available benchmark datasets.
We evaluate text-based (e.g., fastText, Transformers), citation-based (e.g., DeepWalk, Poincar'e), and hybrid methods.
Our experiments show that document representations from averaged fastText word vectors (trained on legal corpora) yield the best results.
arXiv Detail & Related papers (2021-04-28T15:48:19Z) - Learning from Context or Names? An Empirical Study on Neural Relation
Extraction [112.06614505580501]
We study the effect of two main information sources in text: textual context and entity mentions (names)
We propose an entity-masked contrastive pre-training framework for relation extraction (RE)
Our framework can improve the effectiveness and robustness of neural models in different RE scenarios.
arXiv Detail & Related papers (2020-10-05T11:21:59Z) - Learning Neural Textual Representations for Citation Recommendation [7.227232362460348]
We propose a novel approach to citation recommendation using a deep sequential representation of the documents (Sentence-BERT) cascaded with Siamese and triplet networks in a submodular scoring function.
To the best of our knowledge, this is the first approach to combine deep representations and submodular selection for a task of citation recommendation.
arXiv Detail & Related papers (2020-07-08T12:38:50Z) - Towards Making the Most of Context in Neural Machine Translation [112.9845226123306]
We argue that previous research did not make a clear use of the global context.
We propose a new document-level NMT framework that deliberately models the local context of each sentence.
arXiv Detail & Related papers (2020-02-19T03:30:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.