Using Large Language Models to Support Thematic Analysis in Empirical
Legal Studies
- URL: http://arxiv.org/abs/2310.18729v1
- Date: Sat, 28 Oct 2023 15:20:44 GMT
- Title: Using Large Language Models to Support Thematic Analysis in Empirical
Legal Studies
- Authors: Jakub Dr\'apal, Hannes Westermann, Jaromir Savelka
- Abstract summary: We propose a novel framework facilitating effective collaboration of a legal expert with a large language model (LLM)
We employed the framework for an analysis of a dataset (n=785) of facts descriptions from criminal court opinions regarding thefts.
- Score: 0.7673339435080445
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Thematic analysis and other variants of inductive coding are widely used
qualitative analytic methods within empirical legal studies (ELS). We propose a
novel framework facilitating effective collaboration of a legal expert with a
large language model (LLM) for generating initial codes (phase 2 of thematic
analysis), searching for themes (phase 3), and classifying the data in terms of
the themes (to kick-start phase 4). We employed the framework for an analysis
of a dataset (n=785) of facts descriptions from criminal court opinions
regarding thefts. The goal of the analysis was to discover classes of typical
thefts. Our results show that the LLM, namely OpenAI's GPT-4, generated
reasonable initial codes, and it was capable of improving the quality of the
codes based on expert feedback. They also suggest that the model performed well
in zero-shot classification of facts descriptions in terms of the themes.
Finally, the themes autonomously discovered by the LLM appear to map fairly
well to the themes arrived at by legal experts. These findings can be leveraged
by legal researchers to guide their decisions in integrating LLMs into their
thematic analyses, as well as other inductive coding projects.
Related papers
- Interactive Topic Models with Optimal Transport [75.26555710661908]
We present EdTM, as an approach for label name supervised topic modeling.
EdTM models topic modeling as an assignment problem while leveraging LM/LLM based document-topic affinities.
arXiv Detail & Related papers (2024-06-28T13:57:27Z) - CLERC: A Dataset for Legal Case Retrieval and Retrieval-Augmented Analysis Generation [44.67578050648625]
We transform a large open-source legal corpus into a dataset supporting information retrieval (IR) and retrieval-augmented generation (RAG)
This dataset CLERC is constructed for training and evaluating models on their ability to (1) find corresponding citations for a given piece of legal analysis and to (2) compile the text of these citations into a cogent analysis that supports a reasoning goal.
arXiv Detail & Related papers (2024-06-24T23:57:57Z) - Bridging Law and Data: Augmenting Reasoning via a Semi-Structured Dataset with IRAC methodology [22.740895683854568]
This paper introduces LEGALSEMI, a benchmark specifically curated for legal scenario analysis.
LEGALSEMI comprises 54 legal scenarios, each rigorously annotated by legal experts, based on the comprehensive IRAC (Issue, Rule, Application, Conclusion) framework.
A series of experiments were conducted to assess the usefulness of LEGALSEMI for IRAC analysis.
arXiv Detail & Related papers (2024-06-19T04:59:09Z) - Topic Modelling Case Law Using a Large Language Model and a New Taxonomy for UK Law: AI Insights into Summary Judgment [0.0]
This paper develops and applies a novel taxonomy for topic modelling summary judgment cases in the United Kingdom.
Using a curated dataset of summary judgment cases, we use the Large Language Model Claude 3 Opus to explore functional topics and trends.
We find that Claude 3 Opus correctly classified the topic with an accuracy of 87.10%.
arXiv Detail & Related papers (2024-05-21T16:30:25Z) - Large Language Models Offer an Alternative to the Traditional Approach of Topic Modelling [0.9095496510579351]
We investigate the untapped potential of large language models (LLMs) as an alternative for uncovering the underlying topics within extensive text corpora.
Our findings indicate that LLMs with appropriate prompts can stand out as a viable alternative, capable of generating relevant topic titles and adhering to human guidelines to refine and merge topics.
arXiv Detail & Related papers (2024-03-24T17:39:51Z) - LogicAsker: Evaluating and Improving the Logical Reasoning Ability of Large Language Models [63.14196038655506]
We introduce LogicAsker, a novel approach for evaluating and enhancing the logical reasoning capabilities of large language models (LLMs)
Our methodology reveals significant gaps in LLMs' learning of logical rules, with identified reasoning failures ranging from 29% to 90% across different models.
We leverage these findings to construct targeted demonstration examples and fine-tune data, notably enhancing logical reasoning in models like GPT-4o by up to 5%.
arXiv Detail & Related papers (2024-01-01T13:53:53Z) - Can ChatGPT Perform Reasoning Using the IRAC Method in Analyzing Legal
Scenarios Like a Lawyer? [14.103170412148584]
ChatGPT is applied to perform analysis on the corpus using the IRAC method.
Each scenario in the corpus is annotated with a complete IRAC analysis in a semi-structured format.
In addition, we conducted the first empirical assessment of ChatGPT for IRAC analysis in order to understand how well it aligns with the analysis of legal professionals.
arXiv Detail & Related papers (2023-10-23T12:51:49Z) - A Comprehensive Evaluation of Large Language Models on Legal Judgment
Prediction [60.70089334782383]
Large language models (LLMs) have demonstrated great potential for domain-specific applications.
Recent disputes over GPT-4's law evaluation raise questions concerning their performance in real-world legal tasks.
We design practical baseline solutions based on LLMs and test on the task of legal judgment prediction.
arXiv Detail & Related papers (2023-10-18T07:38:04Z) - Sentiment Analysis in the Era of Large Language Models: A Reality Check [69.97942065617664]
This paper investigates the capabilities of large language models (LLMs) in performing various sentiment analysis tasks.
We evaluate performance across 13 tasks on 26 datasets and compare the results against small language models (SLMs) trained on domain-specific datasets.
arXiv Detail & Related papers (2023-05-24T10:45:25Z) - Language Model Decoding as Likelihood-Utility Alignment [54.70547032876017]
We introduce a taxonomy that groups decoding strategies based on their implicit assumptions about how well the model's likelihood is aligned with the task-specific notion of utility.
Specifically, by analyzing the correlation between the likelihood and the utility of predictions across a diverse set of tasks, we provide the first empirical evidence supporting the proposed taxonomy.
arXiv Detail & Related papers (2022-10-13T17:55:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.