Modeling the Diachronic Evolution of Legal Norms: An LRMoo-Based, Component-Level, Event-Centric Approach to Legal Knowledge Graphs
- URL: http://arxiv.org/abs/2506.07853v3
- Date: Tue, 09 Sep 2025 15:15:03 GMT
- Title: Modeling the Diachronic Evolution of Legal Norms: An LRMoo-Based, Component-Level, Event-Centric Approach to Legal Knowledge Graphs
- Authors: Hudson de Martim,
- Abstract summary: This paper proposes a structured, temporal modeling pattern grounded in the LRMoo to address this need.<n>We show how this event-generative architecture enables the precise, deterministic retrieval and reconstruction of any part of a legal text as it existed on a specific date.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Effectively representing legal norms for automated processing is a critical challenge, particularly in tracking the temporal evolution of their hierarchical components. While foundational conceptual frameworks like IFLA LRMoo provide a generic toolkit for bibliographic data, and encoding standards like Akoma Ntoso offer a robust syntax for legal documents, a dedicated, formal modeling pattern for granular, component-level versioning is still required. This limitation hinders the deterministic point-intime reconstruction of legal texts, a fundamental capability for reliable Legal Tech and AI applications. This paper proposes a structured, temporal modeling pattern grounded in the LRMoo ontology to address this need. Our approach models the evolution of a legal norm as a diachronic chain of F2 Expressions. We introduce a key distinction between a language-agnostic Temporal Version (TV)-a semantic snapshot of the norm's structure-and its concrete monolingual realizations, the Language Versions (LV). Both are modeled as F2 Expressions linked by the canonical R76 is derivative of property. This paradigm is applied recursively to the legal text's internal structure, representing it as a parallel hierarchy of abstract Component Works (F1) and their versioned Component Expressions (F2). Furthermore, we formalize the legislative amendment process using the F28 Expression Creation event, allowing changes to be traced from an amending act to its precise effect on the amended norm. Using the Brazilian Federal Constitution as a case study, we demonstrate how this event-centric architecture enables the precise, deterministic retrieval and reconstruction of any part of a legal text as it existed on a specific date. The model provides a robust foundation for building verifiable knowledge graphs and advanced AI tools, overcoming the limitations of current generative models.
Related papers
- IAR2: Improving Autoregressive Visual Generation with Semantic-Detail Associated Token Prediction [77.06211178777939]
IAR2 is an advanced autoregressive framework that enables a hierarchical semantic-detail synthesis process.<n>We show that IAR2 sets a new state-of-the-art for autoregressive image generation, achieving a FID of 1.50 on ImageNet.
arXiv Detail & Related papers (2025-10-08T12:08:21Z) - Legal Requirements Translation from Law [0.0]
We introduce an approach based on textual entailment and in-context learning for automatically generating a canonical representation of legal text.<n>We evaluate our approach on 13 U.S. state data breach notification laws, demonstrating that our generated representations pass approximately 89.4% of test cases and achieve a precision and recall of 82.2 and 88.7, respectively.
arXiv Detail & Related papers (2025-07-03T17:53:48Z) - Respecting Temporal-Causal Consistency: Entity-Event Knowledge Graphs for Retrieval-Augmented Generation [69.45495166424642]
We develop a robust and discriminative QA benchmark to measure temporal, causal, and character consistency understanding in narrative documents.<n>We then introduce Entity-Event RAG (E2RAG), a dual-graph framework that keeps separate entity and event subgraphs linked by a bipartite mapping.<n>Across ChronoQA, our approach outperforms state-of-the-art unstructured and KG-based RAG baselines, with notable gains on causal and character consistency queries.
arXiv Detail & Related papers (2025-06-06T10:07:21Z) - A Foundational Schema.org Mapping for a Legal Knowledge Graph: Representing Brazilian Legal Norms as FRBR Works [0.0]
Structuring legal norms for machine readability is a critical prerequisite for building advanced AI and information retrieval systems.<n>This paper proposes a mapping for the abstract legal Work to the foundational schema.org/Legislation vocabulary.<n>This structured, formal approach provides the essential first step toward creating a deterministic and verifiable knowledge graph.
arXiv Detail & Related papers (2025-05-12T15:11:11Z) - Graph RAG for Legal Norms: A Hierarchical and Temporal Approach [0.0]
This article proposes an adaptation of Graph Retrieval Augmented Generation (Graph RAG) specifically designed for the analysis and comprehension of legal norms.<n>By combining structured knowledge graphs with contextually enriched text segments, Graph RAG offers a promising solution to address the inherent complexity and vast volume of legal data.
arXiv Detail & Related papers (2025-04-29T18:36:57Z) - Transformer-Based Extraction of Statutory Definitions from the U.S. Code [0.0]
We present an advanced NLP system to automatically extract defined terms, their definitions, and their scope from the United States Code (U.S.C.)<n>Our best model achieves 96.8% precision and 98.9% recall (98.2% F1-score)<n>This work contributes to improving accessibility and understanding of legal information while establishing a foundation for downstream legal reasoning tasks.
arXiv Detail & Related papers (2025-04-23T02:09:53Z) - Structured Legal Document Generation in India: A Model-Agnostic Wrapper Approach with VidhikDastaavej [5.790242888372048]
We introduce VidhikDastaavej, a novel, anonymized dataset of private legal documents.<n>We develop NyayaShilp, a fine-tuned legal document generation model specifically adapted to Indian legal texts.
arXiv Detail & Related papers (2025-04-04T14:41:50Z) - WritingBench: A Comprehensive Benchmark for Generative Writing [87.48445972563631]
We present WritingBench, a benchmark designed to evaluate large language models (LLMs) across 6 core writing domains and 100, encompassing creative, persuasive, informative, and technical writing.<n>We propose a query-dependent evaluation framework that empowers LLMs to dynamically generate instance-specific assessment criteria.<n>This framework is complemented by a fine-tuned critic model for criteria-aware scoring, enabling evaluations in style, format and length.
arXiv Detail & Related papers (2025-03-07T08:56:20Z) - Knowledge Graphs Construction from Criminal Court Appeals: Insights from the French Cassation Court [49.1574468325115]
This paper presents a framework for constructing knowledge graphs from appeals to the French Cassation Court.<n>The framework includes a domain-specific ontology and a derived dataset, offering a foundation for structured legal data representation and analysis.
arXiv Detail & Related papers (2025-01-24T15:38:32Z) - Detecting Document-level Paraphrased Machine Generated Content: Mimicking Human Writing Style and Involving Discourse Features [57.34477506004105]
Machine-generated content poses challenges such as academic plagiarism and the spread of misinformation.<n>We introduce novel methodologies and datasets to overcome these challenges.<n>We propose MhBART, an encoder-decoder model designed to emulate human writing style.<n>We also propose DTransformer, a model that integrates discourse analysis through PDTB preprocessing to encode structural features.
arXiv Detail & Related papers (2024-12-17T08:47:41Z) - Physics of Language Models: Part 1, Learning Hierarchical Language Structures [51.68385617116854]
Transformer-based language models are effective but complex, and understanding their inner workings and reasoning mechanisms is a significant challenge.<n>We introduce a family of synthetic CFGs that produce hierarchical rules, capable of generating lengthy sentences.<n>We demonstrate that generative models like GPT can accurately learn and reason over CFG-defined hierarchies and generate sentences based on it.
arXiv Detail & Related papers (2023-05-23T04:28:16Z) - SAILER: Structure-aware Pre-trained Language Model for Legal Case
Retrieval [75.05173891207214]
Legal case retrieval plays a core role in the intelligent legal system.
Most existing language models have difficulty understanding the long-distance dependencies between different structures.
We propose a new Structure-Aware pre-traIned language model for LEgal case Retrieval.
arXiv Detail & Related papers (2023-04-22T10:47:01Z) - The Whole Truth and Nothing But the Truth: Faithful and Controllable
Dialogue Response Generation with Dataflow Transduction and Constrained
Decoding [65.34601470417967]
We describe a hybrid architecture for dialogue response generation that combines the strengths of neural language modeling and rule-based generation.
Our experiments show that this system outperforms both rule-based and learned approaches in human evaluations of fluency, relevance, and truthfulness.
arXiv Detail & Related papers (2022-09-16T09:00:49Z) - Corpus for Automatic Structuring of Legal Documents [1.8025738207124173]
We introduce a corpus of legal judgment documents in English that are segmented into topical and coherent parts.
We develop baseline models for automatically predicting rhetorical roles in a legal document based on the annotated corpus.
We show the application of rhetorical roles to improve performance on the tasks of summarization and legal judgment prediction.
arXiv Detail & Related papers (2022-01-31T11:12:44Z) - DPCL: a Language Template for Normative Specifications [0.0]
Legal core concepts have been proposed to systematize and relationships relevant to reasoning.
No solution amongst those has achieved general acceptance, and no common ground (representational, computational) has been identified.
This presentation will introduce DPCL, a domain-specific language (norms) for specifying higher-level policies.
arXiv Detail & Related papers (2022-01-12T13:51:11Z) - Semantic Segmentation of Legal Documents via Rhetorical Roles [3.285073688021526]
This paper proposes a Rhetorical Roles (RR) system for segmenting a legal document into semantically coherent units.
We develop a multitask learning-based deep learning model with document rhetorical role label shift as an auxiliary task for segmenting a legal document.
arXiv Detail & Related papers (2021-12-03T10:49:19Z) - GRIT: Generative Role-filler Transformers for Document-level Event
Entity Extraction [134.5580003327839]
We introduce a generative transformer-based encoder-decoder framework (GRIT) to model context at the document level.
We evaluate our approach on the MUC-4 dataset, and show that our model performs substantially better than prior work.
arXiv Detail & Related papers (2020-08-21T01:07:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.