Target-aware Abstractive Related Work Generation with Contrastive
Learning
- URL: http://arxiv.org/abs/2205.13339v1
- Date: Thu, 26 May 2022 13:20:51 GMT
- Title: Target-aware Abstractive Related Work Generation with Contrastive
Learning
- Authors: Xiuying Chen, Hind Alamro, Mingzhe Li, Shen Gao, Rui Yan, Xin Gao,
Xiangliang Zhang
- Abstract summary: The related work section is an important component of a scientific paper, which highlights the contribution of the target paper in the context of the reference papers.
Most of the existing related work section generation methods rely on extracting off-the-shelf sentences.
We propose an abstractive target-aware related work generator (TAG), which can generate related work sections consisting of new sentences.
- Score: 48.02845973891943
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The related work section is an important component of a scientific paper,
which highlights the contribution of the target paper in the context of the
reference papers. Authors can save their time and effort by using the
automatically generated related work section as a draft to complete the final
related work. Most of the existing related work section generation methods rely
on extracting off-the-shelf sentences to make a comparative discussion about
the target work and the reference papers. However, such sentences need to be
written in advance and are hard to obtain in practice. Hence, in this paper, we
propose an abstractive target-aware related work generator (TAG), which can
generate related work sections consisting of new sentences. Concretely, we
first propose a target-aware graph encoder, which models the relationships
between reference papers and the target paper with target-centered attention
mechanisms. In the decoding process, we propose a hierarchical decoder that
attends to the nodes of different levels in the graph with keyphrases as
semantic indicators. Finally, to generate a more informative related work, we
propose multi-level contrastive optimization objectives, which aim to maximize
the mutual information between the generated related work with the references
and minimize that with non-references. Extensive experiments on two public
scholar datasets show that the proposed model brings substantial improvements
over several strong baselines in terms of automatic and tailored human
evaluations.
Related papers
- Explaining Relationships Among Research Papers [14.223038413516685]
We propose a feature-based, LLM-prompting approach to generate richer citation texts.
We find a strong correlation between human preference and integrative writing style, suggesting that humans prefer high-level, abstract citations.
arXiv Detail & Related papers (2024-02-20T23:38:39Z) - Leveraging Contextual Information for Effective Entity Salience Detection [21.30389576465761]
We show that fine-tuning medium-sized language models with a cross-encoder style architecture yields substantial performance gains over feature engineering approaches.
We also show that zero-shot prompting of instruction-tuned language models yields inferior results, indicating the task's uniqueness and complexity.
arXiv Detail & Related papers (2023-09-14T19:04:40Z) - Unsupervised Task Graph Generation from Instructional Video Transcripts [53.54435048879365]
We consider a setting where text transcripts of instructional videos performing a real-world activity are provided.
The goal is to identify the key steps relevant to the task as well as the dependency relationship between these key steps.
We propose a novel task graph generation approach that combines the reasoning capabilities of instruction-tuned language models along with clustering and ranking components.
arXiv Detail & Related papers (2023-02-17T22:50:08Z) - CORWA: A Citation-Oriented Related Work Annotation Dataset [4.740962650068886]
In natural language processing, literature reviews are usually conducted under the "Related Work" section.
We train a strong baseline model that automatically tags the CORWA labels on massive unlabeled related work section texts.
We suggest a novel framework for human-in-the-loop, iterative, abstractive related work generation.
arXiv Detail & Related papers (2022-05-07T00:23:46Z) - Revise and Resubmit: An Intertextual Model of Text-based Collaboration
in Peer Review [52.359007622096684]
Peer review is a key component of the publishing process in most fields of science.
Existing NLP studies focus on the analysis of individual texts.
editorial assistance often requires modeling interactions between pairs of texts.
arXiv Detail & Related papers (2022-04-22T16:39:38Z) - Author Clustering and Topic Estimation for Short Texts [69.54017251622211]
We propose a novel model that expands on the Latent Dirichlet Allocation by modeling strong dependence among the words in the same document.
We also simultaneously cluster users, removing the need for post-hoc cluster estimation.
Our method performs as well as -- or better -- than traditional approaches to problems arising in short text.
arXiv Detail & Related papers (2021-06-15T20:55:55Z) - Generating Related Work [37.161925758727456]
We model generating related work sections while being cognisant of the motivation behind citing papers.
Our model outperforms several strong state-of-the-art summarization and multi-document summarization models.
arXiv Detail & Related papers (2021-04-18T00:19:37Z) - Sequential Cross-Document Coreference Resolution [14.099694053823765]
Cross-document coreference resolution is important for the growing interest in multi-document analysis tasks.
We propose a new model that extends the efficient sequential prediction paradigm for coreference resolution to cross-document settings.
Our model incrementally composes mentions into cluster representations and predicts links between a mention and the already constructed clusters.
arXiv Detail & Related papers (2021-04-17T00:46:57Z) - What's New? Summarizing Contributions in Scientific Literature [85.95906677964815]
We introduce a new task of disentangled paper summarization, which seeks to generate separate summaries for the paper contributions and the context of the work.
We extend the S2ORC corpus of academic articles by adding disentangled "contribution" and "context" reference labels.
We propose a comprehensive automatic evaluation protocol which reports the relevance, novelty, and disentanglement of generated outputs.
arXiv Detail & Related papers (2020-11-06T02:23:01Z) - Explaining Relationships Between Scientific Documents [55.23390424044378]
We address the task of explaining relationships between two scientific documents using natural language text.
In this paper we establish a dataset of 622K examples from 154K documents.
arXiv Detail & Related papers (2020-02-02T03:54:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.