Summarization, Simplification, and Generation: The Case of Patents
- URL: http://arxiv.org/abs/2104.14860v1
- Date: Fri, 30 Apr 2021 09:28:29 GMT
- Title: Summarization, Simplification, and Generation: The Case of Patents
- Authors: Silvia Casola and Alberto Lavelli
- Abstract summary: This survey aims at a) describing patents' characteristics and the questions they raise to the current NLP systems, b) critically presenting previous work and its evolution, and c) drawing attention to directions of research in which further work is needed.
To the best of our knowledge, this is the first survey of generative approaches in the patent domain.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We survey Natural Language Processing (NLP) approaches to summarizing,
simplifying, and generating patents' text. While solving these tasks has
important practical applications - given patents' centrality in the R&D process
- patents' idiosyncrasies open peculiar challenges to the current NLP state of
the art. This survey aims at a) describing patents' characteristics and the
questions they raise to the current NLP systems, b) critically presenting
previous work and its evolution, and c) drawing attention to directions of
research in which further work is needed. To the best of our knowledge, this is
the first survey of generative approaches in the patent domain.
Related papers
- Pap2Pat: Towards Automated Paper-to-Patent Drafting using Chunk-based Outline-guided Generation [13.242188189150987]
We present PAP2PAT, a new challenging benchmark of 1.8k patent-paper pairs with document outlines.
Our experiments with current open-weight LLMs and outline-guided generation show that they can effectively use information from the paper but struggle with repetitions, likely due to the inherent repetitiveness of patent language.
arXiv Detail & Related papers (2024-10-09T15:52:48Z) - PatentGPT: A Large Language Model for Patent Drafting Using Knowledge-based Fine-tuning Method [1.4496326701907591]
Existing large language models (LLMs) often fall short in this IP creation domain due to their lack of specialized knowledge and context-awareness.
We propose a groundbreaking framework for Knowledge Fine-Tuning (KFT) of LLMs, designed to endow AI with the ability to autonomously mine, understand, and apply domain-specific knowledge.
Our model, PatentGPT, has demonstrated outstanding performance, scoring up to approximately 400% higher in patent related benchmark tests compared to state-of-the-art models.
arXiv Detail & Related papers (2024-08-26T12:00:29Z) - A Survey on RAG Meeting LLMs: Towards Retrieval-Augmented Large Language Models [71.25225058845324]
Large Language Models (LLMs) have demonstrated revolutionary abilities in language understanding and generation.
Retrieval-Augmented Generation (RAG) can offer reliable and up-to-date external knowledge.
RA-LLMs have emerged to harness external and authoritative knowledge bases, rather than relying on the model's internal knowledge.
arXiv Detail & Related papers (2024-05-10T02:48:45Z) - A Comprehensive Survey on AI-based Methods for Patents [14.090575139188422]
AI-based tools present opportunities to streamline and enhance important tasks in the patent cycle.
This interdisciplinary survey aims to serve as a resource for researchers and practitioners working at the intersection of AI and patent analysis.
arXiv Detail & Related papers (2024-04-02T20:44:06Z) - LLM-based Extraction of Contradictions from Patents [0.0]
This paper goes one step further, as it presents a method to extract TRIZ contradictions from patent texts based on Prompt Engineering.
Our results show that "off-the-shelf" GPT-4 is a serious alternative to existing approaches.
arXiv Detail & Related papers (2024-03-21T09:36:36Z) - Natural Language Processing in Patents: A Survey [0.0]
Patents, encapsulating crucial technical and legal information, present a rich domain for natural language processing (NLP) applications.
As NLP technologies evolve, large language models (LLMs) have demonstrated outstanding capabilities in general text processing and generation tasks.
This paper aims to equip NLP researchers with the essential knowledge to navigate this complex domain efficiently.
arXiv Detail & Related papers (2024-03-06T23:17:16Z) - Report of the 1st Workshop on Generative AI and Law [78.62063815165968]
This report presents the takeaways of the inaugural Workshop on Generative AI and Law (GenLaw)
A cross-disciplinary group of practitioners and scholars from computer science and law convened to discuss the technical, doctrinal, and policy challenges presented by law for Generative AI.
arXiv Detail & Related papers (2023-11-11T04:13:37Z) - Regulation and NLP (RegNLP): Taming Large Language Models [51.41095330188972]
We argue how NLP research can benefit from proximity to regulatory studies and adjacent fields.
We advocate for the development of a new multidisciplinary research space on regulation and NLP.
arXiv Detail & Related papers (2023-10-09T09:22:40Z) - Towards Quantum Federated Learning [80.1976558772771]
Quantum Federated Learning aims to enhance privacy, security, and efficiency in the learning process.
We aim to provide a comprehensive understanding of the principles, techniques, and emerging applications of QFL.
As the field of QFL continues to progress, we can anticipate further breakthroughs and applications across various industries.
arXiv Detail & Related papers (2023-06-16T15:40:21Z) - Evidentiality-guided Generation for Knowledge-Intensive NLP Tasks [59.761411682238645]
Retrieval-augmented generation models have shown state-of-the-art performance across many knowledge-intensive NLP tasks.
We introduce a method to incorporate evidentiality of passages -- whether a passage contains correct evidence to support the output -- into training the generator.
arXiv Detail & Related papers (2021-12-16T08:18:47Z) - Retrieving and Reading: A Comprehensive Survey on Open-domain Question
Answering [62.88322725956294]
We review the latest research trends in OpenQA, with particular attention to systems that incorporate neural MRC techniques.
We introduce modern OpenQA architecture named Retriever-Reader'' and analyze the various systems that follow this architecture.
We then discuss key challenges to developing OpenQA systems and offer an analysis of benchmarks that are commonly used.
arXiv Detail & Related papers (2021-01-04T04:47:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.