Recommending Multiple Positive Citations for Manuscript via
Content-Dependent Modeling and Multi-Positive Triplet
- URL: http://arxiv.org/abs/2111.12899v1
- Date: Thu, 25 Nov 2021 04:09:31 GMT
- Title: Recommending Multiple Positive Citations for Manuscript via
Content-Dependent Modeling and Multi-Positive Triplet
- Authors: Yang Zhang and Qiang Ma
- Abstract summary: We propose a novel scientific paper modeling for citation recommendations, namely Multi-Positive BERT Model for Citation Recommendation (MP-BERT4CR)
The proposed approach has the following advantages: First, the proposed multi-positive objectives are effective to recommend multiple positive candidates.
MP-BERT4CR are also effective in retrieving the full list of co-citations, and historically low-frequent co-citation pairs compared with the prior works.
- Score: 6.7854900381386845
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Considering the rapidly increasing number of academic papers, searching for
and citing appropriate references has become a non-trial task during the wiring
of papers. Recommending a handful of candidate papers to a manuscript before
publication could ease the burden of the authors, and help the reviewers to
check the completeness of the cited resources. Conventional approaches on
citation recommendation generally consider recommending one ground-truth
citation for a query context from an input manuscript, but lack of
consideration on co-citation recommendations. However, a piece of context often
needs to be supported by two or more co-citation pairs. Here, we propose a
novel scientific paper modeling for citation recommendations, namely
Multi-Positive BERT Model for Citation Recommendation (MP-BERT4CR), complied
with a series of Multi-Positive Triplet objectives to recommend multiple
positive citations for a query context. The proposed approach has the following
advantages: First, the proposed multi-positive objectives are effective to
recommend multiple positive candidates. Second, we adopt noise distributions
which are built based on the historical co-citation frequencies, so that
MP-BERT4CR is not only effective on recommending high-frequent co-citation
pairs; but also the performances on retrieving the low-frequent ones are
significantly improved. Third, we propose a dynamic context sampling strategy
which captures the ``macro-scoped'' citing intents from a manuscript and
empowers the citation embeddings to be content-dependent, which allow the
algorithm to further improve the performances. Single and multiple positive
recommendation experiments testified that MP-BERT4CR delivered significant
improvements. In addition, MP-BERT4CR are also effective in retrieving the full
list of co-citations, and historically low-frequent co-citation pairs compared
with the prior works.
Related papers
- JudgeRank: Leveraging Large Language Models for Reasoning-Intensive Reranking [81.88787401178378]
We introduce JudgeRank, a novel agentic reranker that emulates human cognitive processes when assessing document relevance.
We evaluate JudgeRank on the reasoning-intensive BRIGHT benchmark, demonstrating substantial performance improvements over first-stage retrieval methods.
In addition, JudgeRank performs on par with fine-tuned state-of-the-art rerankers on the popular BEIR benchmark, validating its zero-shot generalization capability.
arXiv Detail & Related papers (2024-10-31T18:43:12Z) - Ground Every Sentence: Improving Retrieval-Augmented LLMs with Interleaved Reference-Claim Generation [51.8188846284153]
RAG has been widely adopted to enhance Large Language Models (LLMs)
Attributed Text Generation (ATG) has attracted growing attention, which provides citations to support the model's responses in RAG.
This paper proposes a fine-grained ATG method called ReClaim(Refer & Claim), which alternates the generation of references and answers step by step.
arXiv Detail & Related papers (2024-07-01T20:47:47Z) - ALiiCE: Evaluating Positional Fine-grained Citation Generation [54.19617927314975]
We propose ALiiCE, the first automatic evaluation framework for fine-grained citation generation.
Our framework first parses the sentence claim into atomic claims via dependency analysis and then calculates citation quality at the atomic claim level.
We evaluate the positional fine-grained citation generation performance of several Large Language Models on two long-form QA datasets.
arXiv Detail & Related papers (2024-06-19T09:16:14Z) - ILCiteR: Evidence-grounded Interpretable Local Citation Recommendation [31.259805200946175]
We introduce the evidence-grounded local citation recommendation task, where the target latent space comprises evidence spans for recommending specific papers.
Unlike past formulations that simply output recommendations, ILCiteR retrieves ranked lists of evidence span and recommended paper pairs.
We contribute a novel dataset for the evidence-grounded local citation recommendation task and demonstrate the efficacy of our proposed conditional neural rank-ensembling approach for re-ranking evidence spans.
arXiv Detail & Related papers (2024-03-13T17:38:05Z) - Tag-Aware Document Representation for Research Paper Recommendation [68.8204255655161]
We propose a hybrid approach that leverages deep semantic representation of research papers based on social tags assigned by users.
The proposed model is effective in recommending research papers even when the rating data is very sparse.
arXiv Detail & Related papers (2022-09-08T09:13:07Z) - QuoteR: A Benchmark of Quote Recommendation for Writing [80.83859760380616]
We build a large and fully open quote recommendation dataset called QuoteR.
We conduct an extensive evaluation of existing quote recommendation methods on QuoteR.
We propose a new quote recommendation model that significantly outperforms previous methods on all three parts of QuoteR.
arXiv Detail & Related papers (2022-02-26T14:01:44Z) - Learning Neural Textual Representations for Citation Recommendation [7.227232362460348]
We propose a novel approach to citation recommendation using a deep sequential representation of the documents (Sentence-BERT) cascaded with Siamese and triplet networks in a submodular scoring function.
To the best of our knowledge, this is the first approach to combine deep representations and submodular selection for a task of citation recommendation.
arXiv Detail & Related papers (2020-07-08T12:38:50Z) - Context-Based Quotation Recommendation [60.93257124507105]
We propose a novel context-aware quote recommendation system.
It generates a ranked list of quotable paragraphs and spans of tokens from a given source document.
We conduct experiments on a collection of speech transcripts and associated news articles.
arXiv Detail & Related papers (2020-05-17T17:49:53Z) - HybridCite: A Hybrid Model for Context-Aware Citation Recommendation [0.0]
We develop citation recommendation approaches based on embeddings, topic modeling, and information retrieval techniques.
We combine, for the first time to the best of our knowledge, the best-performing algorithms into a semi-genetic hybrid recommender system.
Our evaluation results show that a hybrid model containing embedding and information retrieval-based components outperforms its individual components and further algorithms by a large margin.
arXiv Detail & Related papers (2020-02-15T16:19:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.