When Large Language Models Meet Citation: A Survey
- URL: http://arxiv.org/abs/2309.09727v1
- Date: Mon, 18 Sep 2023 12:48:48 GMT
- Title: When Large Language Models Meet Citation: A Survey
- Authors: Yang Zhang, Yufei Wang, Kai Wang, Quan Z. Sheng, Lina Yao, Adnan
Mahmood, Wei Emma Zhang and Rongying Zhao
- Abstract summary: Large Language Models (LLMs) could be helpful in capturing fine-grained citation information via the corresponding textual context.
Citations also establish connections among scientific papers, providing high-quality inter-document relationships.
We review the application of LLMs for in-text citation analysis tasks, including citation classification, citation-based summarization, and citation recommendation.
- Score: 37.01594297337486
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Citations in scholarly work serve the essential purpose of acknowledging and
crediting the original sources of knowledge that have been incorporated or
referenced. Depending on their surrounding textual context, these citations are
used for different motivations and purposes. Large Language Models (LLMs) could
be helpful in capturing these fine-grained citation information via the
corresponding textual context, thereby enabling a better understanding towards
the literature. Furthermore, these citations also establish connections among
scientific papers, providing high-quality inter-document relationships and
human-constructed knowledge. Such information could be incorporated into LLMs
pre-training and improve the text representation in LLMs. Therefore, in this
paper, we offer a preliminary review of the mutually beneficial relationship
between LLMs and citation analysis. Specifically, we review the application of
LLMs for in-text citation analysis tasks, including citation classification,
citation-based summarization, and citation recommendation. We then summarize
the research pertinent to leveraging citation linkage knowledge to improve text
representations of LLMs via citation prediction, network structure information,
and inter-document relationship. We finally provide an overview of these
contemporary methods and put forth potential promising avenues in combining
LLMs and citation analysis for further investigation.
Related papers
- Ground Every Sentence: Improving Retrieval-Augmented LLMs with Interleaved Reference-Claim Generation [51.8188846284153]
RAG has been widely adopted to enhance Large Language Models (LLMs)
Attributed Text Generation (ATG) has attracted growing attention, which provides citations to support the model's responses in RAG.
This paper proposes a fine-grained ATG method called ReClaim(Refer & Claim), which alternates the generation of references and answers step by step.
arXiv Detail & Related papers (2024-07-01T20:47:47Z) - Verifiable Generation with Subsentence-Level Fine-Grained Citations [13.931548733211436]
Verifiable generation requires large language models to cite source documents supporting their outputs.
Previous work mainly targets the generation of sentence-level citations, lacking specificity about which parts of a sentence are backed by the cited sources.
This work studies verifiable generation with subsentence-level fine-grained citations for more precise location of generated content supported by the cited sources.
arXiv Detail & Related papers (2024-06-10T09:32:37Z) - Context-Enhanced Language Models for Generating Multi-Paper Citations [35.80247519023821]
We propose a method that leverages Large Language Models (LLMs) to generate multi-citation sentences.
Our approach involves a single source paper and a collection of target papers, culminating in a coherent paragraph containing multi-sentence citation text.
arXiv Detail & Related papers (2024-04-22T04:30:36Z) - Source-Aware Training Enables Knowledge Attribution in Language Models [81.13048060332775]
Large language models (LLMs) learn a vast amount of knowledge during pretraining, but they are often oblivious to the source(s) of such knowledge.
We investigate the problem of intrinsic source citation, where LLMs are required to cite the pretraining source supporting a generated response.
Our training recipe can enable faithful attribution to the pretraining data without a substantial impact on the model's quality compared to standard pretraining.
arXiv Detail & Related papers (2024-04-01T09:39:38Z) - Generative Context-aware Fine-tuning of Self-supervised Speech Models [54.389711404209415]
We study the use of generative large language models (LLM) generated context information.
We propose an approach to distill the generated information during fine-tuning of self-supervised speech models.
We evaluate the proposed approach using the SLUE and Libri-light benchmarks for several downstream tasks: automatic speech recognition, named entity recognition, and sentiment analysis.
arXiv Detail & Related papers (2023-12-15T15:46:02Z) - Effective Large Language Model Adaptation for Improved Grounding and Citation Generation [48.07830615309543]
This paper focuses on improving large language models (LLMs) by grounding their responses in retrieved passages and by providing citations.
We propose a new framework, AGREE, that improves the grounding from a holistic perspective.
Our framework tunes LLMs to selfground the claims in their responses and provide accurate citations to retrieved documents.
arXiv Detail & Related papers (2023-11-16T03:22:25Z) - Citation: A Key to Building Responsible and Accountable Large Language Models [25.671237896575693]
Large Language Models (LLMs) bring transformative benefits alongside unique challenges, including intellectual property (IP) and ethical concerns.
This position paper explores a novel angle to mitigate these risks, drawing parallels between LLMs and established web systems.
arXiv Detail & Related papers (2023-07-05T10:25:45Z) - Enabling Large Language Models to Generate Text with Citations [37.64884969997378]
Large language models (LLMs) have emerged as a widely-used tool for information seeking.
Our aim is to allow LLMs to generate text with citations, improving their factual correctness and verifiability.
We propose ALCE, the first benchmark for Automatic LLMs' Citation Evaluation.
arXiv Detail & Related papers (2023-05-24T01:53:49Z) - Inline Citation Classification using Peripheral Context and
Time-evolving Augmentation [23.88211560188731]
We propose a new dataset, named 3Cext, which provides discourse information using the cited sentences.
We propose PeriCite, a Transformer-based deep neural network that fuses peripheral sentences and domain knowledge.
arXiv Detail & Related papers (2023-03-01T09:11:07Z) - CitationIE: Leveraging the Citation Graph for Scientific Information
Extraction [89.33938657493765]
We use the citation graph of referential links between citing and cited papers.
We observe a sizable improvement in end-to-end information extraction over the state-of-the-art.
arXiv Detail & Related papers (2021-06-03T03:00:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.