Towards generating citation sentences for multiple references with
intent control
- URL: http://arxiv.org/abs/2112.01332v1
- Date: Thu, 2 Dec 2021 15:32:24 GMT
- Title: Towards generating citation sentences for multiple references with
intent control
- Authors: Jia-Yan Wu, Alexander Te-Wei Shieh, Shih-Ju Hsu, Yun-Nung Chen
- Abstract summary: We build a novel generation model with the Fusion-in-Decoder approach to cope with multiple long inputs.
Experiments demonstrate that the proposed approaches provide much more comprehensive features for generating citation sentences.
- Score: 86.53829532976303
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine-generated citation sentences can aid automated scientific literature
review and assist article writing. Current methods in generating citation text
were limited to single citation generation using the citing document and a
cited document as input. However, in real-world situations, writers often
summarize several studies in one sentence or discuss relevant information
across the entire paragraph. In addition, multiple citation intents have been
previously identified, implying that writers may need control over the intents
of generated sentences to cover different scenarios. Therefore, this work
focuses on generating multiple citations and releasing a newly collected
dataset named CiteMI to drive the future research. We first build a novel
generation model with the Fusion-in-Decoder approach to cope with multiple long
inputs. Second, we incorporate the predicted citation intents into training for
intent control. The experiments demonstrate that the proposed approaches
provide much more comprehensive features for generating citation sentences.
Related papers
- ALiiCE: Evaluating Positional Fine-grained Citation Generation [54.19617927314975]
We propose ALiiCE, the first automatic evaluation framework for fine-grained citation generation.
Our framework first parses the sentence claim into atomic claims via dependency analysis and then calculates citation quality at the atomic claim level.
We evaluate the positional fine-grained citation generation performance of several Large Language Models on two long-form QA datasets.
arXiv Detail & Related papers (2024-06-19T09:16:14Z) - Context-Enhanced Language Models for Generating Multi-Paper Citations [35.80247519023821]
We propose a method that leverages Large Language Models (LLMs) to generate multi-citation sentences.
Our approach involves a single source paper and a collection of target papers, culminating in a coherent paragraph containing multi-sentence citation text.
arXiv Detail & Related papers (2024-04-22T04:30:36Z) - Contextualizing Generated Citation Texts [11.531517736126657]
We propose a simple modification to the citation text generation task.
The generation target is not only the citation itself, but the entire context window, including the target citation.
arXiv Detail & Related papers (2024-02-28T05:24:21Z) - CiteCaseLAW: Citation Worthiness Detection in Caselaw for Legal
Assistive Writing [44.75251805925605]
We introduce a labeled dataset of 178M sentences for citation-worthiness detection in the legal domain from the Caselaw Access Project (CAP)
The performance of various deep learning models was examined on this novel dataset.
The domain-specific pre-trained model tends to outperform other models, with an 88% F1-score for the citation-worthiness detection task.
arXiv Detail & Related papers (2023-05-03T04:20:56Z) - CiteBench: A benchmark for Scientific Citation Text Generation [69.37571393032026]
CiteBench is a benchmark for citation text generation.
We make the code for CiteBench publicly available at https://github.com/UKPLab/citebench.
arXiv Detail & Related papers (2022-12-19T16:10:56Z) - Controllable Citation Sentence Generation with Language Models [11.186252009101077]
We propose to integrate the manuscript context, the context of the referenced paper, and the desired control attributes into a structured template and use it to fine-tune a language model (LM) via next-token prediction.
The proposed workflow harmoniously combines citation attribute suggestion and conditional citation generation into one LM, allowing for better user control.
arXiv Detail & Related papers (2022-11-14T01:54:08Z) - Deep Graph Learning for Anomalous Citation Detection [55.81334139806342]
We propose a novel deep graph learning model, namely GLAD (Graph Learning for Anomaly Detection), to identify anomalies in citation networks.
Within the GLAD framework, we propose an algorithm called CPU (Citation PUrpose) to discover the purpose of citation based on citation texts.
arXiv Detail & Related papers (2022-02-23T09:05:28Z) - MultiCite: Modeling realistic citations requires moving beyond the
single-sentence single-label setting [13.493267499658527]
We release MultiCite, a new dataset of 12,653 citation contexts from over 1,200 computational linguistics papers.
We show how our dataset, while still usable for training classic CCA models, also supports the development of new types of models for CCA beyond fixed-width text classification.
arXiv Detail & Related papers (2021-07-01T12:54:23Z) - Enhancing Scientific Papers Summarization with Citation Graph [78.65955304229863]
We redefine the task of scientific papers summarization by utilizing their citation graph.
We construct a novel scientific papers summarization dataset Semantic Scholar Network (SSN) which contains 141K research papers in different domains.
Our model can achieve competitive performance when compared with the pretrained models.
arXiv Detail & Related papers (2021-04-07T11:13:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.