Contextualizing Generated Citation Texts
- URL: http://arxiv.org/abs/2402.18054v1
- Date: Wed, 28 Feb 2024 05:24:21 GMT
- Title: Contextualizing Generated Citation Texts
- Authors: Biswadip Mandal, Xiangci Li, Jessica Ouyang
- Abstract summary: We propose a simple modification to the citation text generation task.
The generation target is not only the citation itself, but the entire context window, including the target citation.
- Score: 11.531517736126657
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Abstractive citation text generation is usually framed as an infilling task,
where a sequence-to-sequence model is trained to generate a citation given a
reference paper and the context window around the target; the generated
citation should be a brief discussion of the reference paper as it relates to
the citing context. However, examining a recent LED-based citation generation
system, we find that many of the generated citations are generic summaries of
the reference papers main contribution, ignoring the citation contexts focus on
a different topic. To address this problem, we propose a simple modification to
the citation text generation task: the generation target is not only the
citation itself, but the entire context window, including the target citation.
This approach can be easily applied to any abstractive citation generation
system, and our experimental results show that training in this way is
preferred by human readers and allows the generation model to make use of
contextual clues about what topic to discuss and what stance to take.
Related papers
- SelfCite: Self-Supervised Alignment for Context Attribution in Large Language Models [51.90867482317985]
SelfCite is a self-supervised approach that aligns LLMs to generate high-quality, fine-grained, sentence-level citations for statements in generated responses.
Instead of relying on costly and labor-intensive annotations, SelfCite leverages a reward signal provided by the LLM itself through context ablation.
The effectiveness of SelfCite is demonstrated by increasing citation F1 up to 5.3 points on the LongBench-Cite benchmark across five long-form question answering tasks.
arXiv Detail & Related papers (2025-02-13T18:55:13Z) - Citation Recommendation based on Argumentative Zoning of User Queries [7.596930973436683]
argumentative zoning is to identify the argumentative and rhetorical structure in scientific literature.
In this paper, a multi-task learning model is built for citation recommendation and argumentative zoning classification.
arXiv Detail & Related papers (2025-01-30T12:08:00Z) - Ground Every Sentence: Improving Retrieval-Augmented LLMs with Interleaved Reference-Claim Generation [51.8188846284153]
RAG has been widely adopted to enhance Large Language Models (LLMs)
Attributed Text Generation (ATG) has attracted growing attention, which provides citations to support the model's responses in RAG.
This paper proposes a fine-grained ATG method called ReClaim(Refer & Claim), which alternates the generation of references and answers step by step.
arXiv Detail & Related papers (2024-07-01T20:47:47Z) - ALiiCE: Evaluating Positional Fine-grained Citation Generation [54.19617927314975]
We propose ALiiCE, the first automatic evaluation framework for fine-grained citation generation.
Our framework first parses the sentence claim into atomic claims via dependency analysis and then calculates citation quality at the atomic claim level.
We evaluate the positional fine-grained citation generation performance of several Large Language Models on two long-form QA datasets.
arXiv Detail & Related papers (2024-06-19T09:16:14Z) - Context-Enhanced Language Models for Generating Multi-Paper Citations [35.80247519023821]
We propose a method that leverages Large Language Models (LLMs) to generate multi-citation sentences.
Our approach involves a single source paper and a collection of target papers, culminating in a coherent paragraph containing multi-sentence citation text.
arXiv Detail & Related papers (2024-04-22T04:30:36Z) - CiteBench: A benchmark for Scientific Citation Text Generation [69.37571393032026]
CiteBench is a benchmark for citation text generation.
We make the code for CiteBench publicly available at https://github.com/UKPLab/citebench.
arXiv Detail & Related papers (2022-12-19T16:10:56Z) - Controllable Citation Sentence Generation with Language Models [11.186252009101077]
We propose to integrate the manuscript context, the context of the referenced paper, and the desired control attributes into a structured template and use it to fine-tune a language model (LM) via next-token prediction.
The proposed workflow harmoniously combines citation attribute suggestion and conditional citation generation into one LM, allowing for better user control.
arXiv Detail & Related papers (2022-11-14T01:54:08Z) - Towards generating citation sentences for multiple references with
intent control [86.53829532976303]
We build a novel generation model with the Fusion-in-Decoder approach to cope with multiple long inputs.
Experiments demonstrate that the proposed approaches provide much more comprehensive features for generating citation sentences.
arXiv Detail & Related papers (2021-12-02T15:32:24Z) - Enhancing Scientific Papers Summarization with Citation Graph [78.65955304229863]
We redefine the task of scientific papers summarization by utilizing their citation graph.
We construct a novel scientific papers summarization dataset Semantic Scholar Network (SSN) which contains 141K research papers in different domains.
Our model can achieve competitive performance when compared with the pretrained models.
arXiv Detail & Related papers (2021-04-07T11:13:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.