Metric-guided Distillation: Distilling Knowledge from the Metric to
Ranker and Retriever for Generative Commonsense Reasoning
- URL: http://arxiv.org/abs/2210.11708v1
- Date: Fri, 21 Oct 2022 03:34:24 GMT
- Title: Metric-guided Distillation: Distilling Knowledge from the Metric to
Ranker and Retriever for Generative Commonsense Reasoning
- Authors: Xingwei He, Yeyun Gong, A-Long Jin, Weizhen Qi, Hang Zhang, Jian Jiao,
Bartuer Zhou, Biao Cheng, SM Yiu and Nan Duan
- Abstract summary: We propose a metric distillation rule to distill knowledge from the metric to the ranker.
We further transfer the critical knowledge summarized by the distilled ranker to the retriever.
Experimental results on the CommonGen benchmark verify the effectiveness of our proposed method.
- Score: 48.18060169551869
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Commonsense generation aims to generate a realistic sentence describing a
daily scene under the given concepts, which is very challenging, since it
requires models to have relational reasoning and compositional generalization
capabilities. Previous work focuses on retrieving prototype sentences for the
provided concepts to assist generation. They first use a sparse retriever to
retrieve candidate sentences, then re-rank the candidates with a ranker.
However, the candidates returned by their ranker may not be the most relevant
sentences, since the ranker treats all candidates equally without considering
their relevance to the reference sentences of the given concepts. Another
problem is that re-ranking is very expensive, but only using retrievers will
seriously degrade the performance of their generation models. To solve these
problems, we propose the metric distillation rule to distill knowledge from the
metric (e.g., BLEU) to the ranker. We further transfer the critical knowledge
summarized by the distilled ranker to the retriever. In this way, the relevance
scores of candidate sentences predicted by the ranker and retriever will be
more consistent with their quality measured by the metric. Experimental results
on the CommonGen benchmark verify the effectiveness of our proposed method: (1)
Our generation model with the distilled ranker achieves a new state-of-the-art
result. (2) Our generation model with the distilled retriever even surpasses
the previous SOTA.
Related papers
- Learning to Rank in Generative Retrieval [62.91492903161522]
Generative retrieval aims to generate identifier strings of relevant passages as the retrieval target.
We propose a learning-to-rank framework for generative retrieval, dubbed LTRGR.
This framework only requires an additional learning-to-rank training phase to enhance current generative retrieval systems.
arXiv Detail & Related papers (2023-06-27T05:48:14Z) - ReFIT: Relevance Feedback from a Reranker during Inference [109.33278799999582]
Retrieve-and-rerank is a prevalent framework in neural information retrieval.
We propose to leverage the reranker to improve recall by making it provide relevance feedback to the retriever at inference time.
arXiv Detail & Related papers (2023-05-19T15:30:33Z) - LED: Lexicon-Enlightened Dense Retriever for Large-Scale Retrieval [68.85686621130111]
We propose to make a dense retriever align a well-performing lexicon-aware representation model.
We evaluate our model on three public benchmarks, which shows that with a comparable lexicon-aware retriever as the teacher, our proposed dense model can bring consistent and significant improvements.
arXiv Detail & Related papers (2022-08-29T15:09:28Z) - Towards Robust Ranker for Text Retrieval [83.15191578888188]
A ranker plays an indispensable role in the de facto'retrieval & rerank' pipeline.
A ranker plays an indispensable role in the de facto'retrieval & rerank' pipeline.
arXiv Detail & Related papers (2022-06-16T10:27:46Z) - Adversarial Retriever-Ranker for dense text retrieval [51.87158529880056]
We present Adversarial Retriever-Ranker (AR2), which consists of a dual-encoder retriever plus a cross-encoder ranker.
AR2 consistently and significantly outperforms existing dense retriever methods.
This includes the improvements on Natural Questions R@5 to 77.9%(+2.1%), TriviaQA R@5 to 78.2%(+1.4), and MS-MARCO MRR@10 to 39.5%(+1.3%)
arXiv Detail & Related papers (2021-10-07T16:41:15Z) - Retrieval Enhanced Model for Commonsense Generation [27.808363395849536]
We propose a novel framework using retrieval methods to enhance both the pre-training and fine-tuning for commonsense generation.
We retrieve prototype sentence candidates by concept matching and use them as auxiliary input.
We demonstrate experimentally on the large-scale CommonGen benchmark that our approach achieves new state-of-the-art results.
arXiv Detail & Related papers (2021-05-24T09:49:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.