Distill-VQ: Learning Retrieval Oriented Vector Quantization By
Distilling Knowledge from Dense Embeddings
- URL: http://arxiv.org/abs/2204.00185v1
- Date: Fri, 1 Apr 2022 03:30:40 GMT
- Title: Distill-VQ: Learning Retrieval Oriented Vector Quantization By
Distilling Knowledge from Dense Embeddings
- Authors: Shitao Xiao, Zheng Liu, Weihao Han, Jianjin Zhang, Defu Lian, Yeyun
Gong, Qi Chen, Fan Yang, Hao Sun, Yingxia Shao, Denvy Deng, Qi Zhang, Xing
Xie
- Abstract summary: We propose Distill-VQ, which unifies the learning of IVF and PQ within a knowledge distillation framework.
The VQ modules are treated as the "students", which are learned to reproduce the predicted relevance.
By doing so, Distill-VQ is able to derive substantial training signals from the massive unlabeled data.
- Score: 50.903547893204234
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Vector quantization (VQ) based ANN indexes, such as Inverted File System
(IVF) and Product Quantization (PQ), have been widely applied to embedding
based document retrieval thanks to the competitive time and memory efficiency.
Originally, VQ is learned to minimize the reconstruction loss, i.e., the
distortions between the original dense embeddings and the reconstructed
embeddings after quantization. Unfortunately, such an objective is inconsistent
with the goal of selecting ground-truth documents for the input query, which
may cause severe loss of retrieval quality. Recent works identify such a
defect, and propose to minimize the retrieval loss through contrastive
learning. However, these methods intensively rely on queries with ground-truth
documents, whose performance is limited by the insufficiency of labeled data.
In this paper, we propose Distill-VQ, which unifies the learning of IVF and
PQ within a knowledge distillation framework. In Distill-VQ, the dense
embeddings are leveraged as "teachers", which predict the query's relevance to
the sampled documents. The VQ modules are treated as the "students", which are
learned to reproduce the predicted relevance, such that the reconstructed
embeddings may fully preserve the retrieval result of the dense embeddings. By
doing so, Distill-VQ is able to derive substantial training signals from the
massive unlabeled data, which significantly contributes to the retrieval
quality. We perform comprehensive explorations for the optimal conduct of
knowledge distillation, which may provide useful insights for the learning of
VQ based ANN index. We also experimentally show that the labeled data is no
longer a necessity for high-quality vector quantization, which indicates
Distill-VQ's strong applicability in practice.
Related papers
- Learning to Compress Contexts for Efficient Knowledge-based Visual Question Answering [44.54319663913782]
We propose Retrieval-Augmented MLLM with Compressed Contexts (RACC)
RACC achieves a state-of-the-art (SOTA) performance of 62.9% on OK-VQA.
It significantly reduces inference latency by 22.0%-59.7% compared to the prominent RAVQA-v2.
arXiv Detail & Related papers (2024-09-11T15:11:39Z) - HyperVQ: MLR-based Vector Quantization in Hyperbolic Space [56.4245885674567]
We study the use of hyperbolic spaces for vector quantization (HyperVQ)
We show that hyperVQ performs comparably in reconstruction and generative tasks while outperforming VQ in discriminative tasks and learning a highly disentangled latent space.
arXiv Detail & Related papers (2024-03-18T03:17:08Z) - LL-VQ-VAE: Learnable Lattice Vector-Quantization For Efficient
Representations [0.0]
We introduce learnable lattice vector quantization and demonstrate its effectiveness for learning discrete representations.
Our method, termed LL-VQ-VAE, replaces the vector quantization layer in VQ-VAE with lattice-based discretization.
Compared to VQ-VAE, our method obtains lower reconstruction errors under the same training conditions, trains in a fraction of the time, and with a constant number of parameters.
arXiv Detail & Related papers (2023-10-13T20:03:18Z) - Noise-Robust Dense Retrieval via Contrastive Alignment Post Training [89.29256833403167]
Contrastive Alignment POst Training (CAPOT) is a highly efficient finetuning method that improves model robustness without requiring index regeneration.
CAPOT enables robust retrieval by freezing the document encoder while the query encoder learns to align noisy queries with their unaltered root.
We evaluate CAPOT noisy variants of MSMARCO, Natural Questions, and Trivia QA passage retrieval, finding CAPOT has a similar impact as data augmentation with none of its overhead.
arXiv Detail & Related papers (2023-04-06T22:16:53Z) - A Unified End-to-End Retriever-Reader Framework for Knowledge-based VQA [67.75989848202343]
This paper presents a unified end-to-end retriever-reader framework towards knowledge-based VQA.
We shed light on the multi-modal implicit knowledge from vision-language pre-training models to mine its potential in knowledge reasoning.
Our scheme is able to not only provide guidance for knowledge retrieval, but also drop these instances potentially error-prone towards question answering.
arXiv Detail & Related papers (2022-06-30T02:35:04Z) - CONVIQT: Contrastive Video Quality Estimator [63.749184706461826]
Perceptual video quality assessment (VQA) is an integral component of many streaming and video sharing platforms.
Here we consider the problem of learning perceptually relevant video quality representations in a self-supervised manner.
Our results indicate that compelling representations with perceptual bearing can be obtained using self-supervised learning.
arXiv Detail & Related papers (2022-06-29T15:22:01Z) - Questions Are All You Need to Train a Dense Passage Retriever [123.13872383489172]
ART is a new corpus-level autoencoding approach for training dense retrieval models that does not require any labeled training data.
It uses a new document-retrieval autoencoding scheme, where (1) an input question is used to retrieve a set of evidence documents, and (2) the documents are then used to compute the probability of reconstructing the original question.
arXiv Detail & Related papers (2022-06-21T18:16:31Z) - Q-Learning with Differential Entropy of Q-Tables [4.221871357181261]
We conjecture that the reduction in performance during prolonged training sessions of Q-learning is caused by a loss of information.
We introduce Differential Entropy of Q-tables (DE-QT) as an external information loss detector to the Q-learning algorithm.
arXiv Detail & Related papers (2020-06-26T04:37:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.