Hot-Refresh Model Upgrades with Regression-Alleviating Compatible
Training in Image Retrieval
- URL: http://arxiv.org/abs/2201.09724v1
- Date: Mon, 24 Jan 2022 14:59:12 GMT
- Title: Hot-Refresh Model Upgrades with Regression-Alleviating Compatible
Training in Image Retrieval
- Authors: Binjie Zhang, Yixiao Ge, Yantao Shen, Yu Li, Chun Yuan, Xuyuan Xu,
Yexin Wang, Ying Shan
- Abstract summary: cold-refresh model upgrades can only deploy new models after the gallery is overall backfilled, taking weeks or even months for massive data.
In contrast, hot-refresh model upgrades deploy the new model immediately and then gradually improve the retrieval accuracy by backfilling the gallery on-the-fly.
- Score: 34.84329831602699
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The task of hot-refresh model upgrades of image retrieval systems plays an
essential role in the industry but has never been investigated in academia
before. Conventional cold-refresh model upgrades can only deploy new models
after the gallery is overall backfilled, taking weeks or even months for
massive data. In contrast, hot-refresh model upgrades deploy the new model
immediately and then gradually improve the retrieval accuracy by backfilling
the gallery on-the-fly. Compatible training has made it possible, however, the
problem of model regression with negative flips poses a great challenge to the
stable improvement of user experience. We argue that it is mainly due to the
fact that new-to-old positive query-gallery pairs may show less similarity than
new-to-new negative pairs. To solve the problem, we introduce a
Regression-Alleviating Compatible Training (RACT) method to properly constrain
the feature compatibility while reducing negative flips. The core is to
encourage the new-to-old positive pairs to be more similar than both the
new-to-old negative pairs and the new-to-new negative pairs. An efficient
uncertainty-based backfilling strategy is further introduced to fasten accuracy
improvements. Extensive experiments on large-scale retrieval benchmarks (e.g.,
Google Landmark) demonstrate that our RACT effectively alleviates the model
regression for one more step towards seamless model upgrades. The code will be
available at https://github.com/binjiezhang/RACT_ICLR2022.
Related papers
- Backward-Compatible Aligned Representations via an Orthogonal Transformation Layer [20.96380700548786]
Visual retrieval systems face challenges when updating models with improved representations due to misalignment between the old and new representations.
Prior research has explored backward-compatible training methods that enable direct comparisons between new and old representations without backfilling.
In this paper, we address achieving a balance between backward compatibility and the performance of independently trained models.
arXiv Detail & Related papers (2024-08-16T15:05:28Z) - MixBCT: Towards Self-Adapting Backward-Compatible Training [66.52766344751635]
We propose MixBCT, a simple yet highly effective backward-compatible training method.
We conduct experiments on the large-scale face recognition datasets MS1Mv3 and IJB-C.
arXiv Detail & Related papers (2023-08-14T05:55:38Z) - FastFill: Efficient Compatible Model Update [40.27741553705222]
FastFill is a compatible model update process using feature alignment and policy based partial backfilling.
We show that previous backfilling strategies suffer from decreased performance and demonstrate the importance of both the training objective and the ordering in online partial backfilling.
arXiv Detail & Related papers (2023-03-08T18:03:51Z) - Improving Prediction Backward-Compatiblility in NLP Model Upgrade with
Gated Fusion [8.173078054056337]
When upgrading neural models to a newer version, new errors that were not encountered in the legacy version can be introduced, known as regression errors.
We propose a novel method, Gated Fusion, that promotes backward compatibility via learning to mix predictions between old and new models.
arXiv Detail & Related papers (2023-02-04T03:40:35Z) - Backward Compatibility During Data Updates by Weight Interpolation [17.502410289568587]
We study the problem of regression during data updates and propose Backward Compatible Weight Interpolation (BCWI)
BCWI reduces negative flips without sacrificing the improved accuracy of the new model.
We also explore the use of importance weighting during and averaging the weights of multiple new models in order to further reduce negative flips.
arXiv Detail & Related papers (2023-01-25T12:23:10Z) - Online Backfilling with No Regret for Large-Scale Image Retrieval [50.162438586686356]
Backfilling is the process of re-extracting all gallery embeddings from upgraded models in image retrieval systems.
We propose an online backfilling algorithm, which enables us to achieve a progressive performance improvement during the backfilling process.
We incorporate a reverse transformation module for more effective and efficient merging, which is further enhanced by adopting a metric-compatible contrastive learning approach.
arXiv Detail & Related papers (2023-01-10T03:10:32Z) - Forward Compatible Training for Representation Learning [53.300192863727226]
backward compatible training (BCT) modifies training of the new model to make its representations compatible with those of the old model.
BCT can significantly hinder the performance of the new model.
In this work, we propose a new learning paradigm for representation learning: forward compatible training (FCT)
arXiv Detail & Related papers (2021-12-06T06:18:54Z) - Regression Bugs Are In Your Model! Measuring, Reducing and Analyzing
Regressions In NLP Model Updates [68.09049111171862]
This work focuses on quantifying, reducing and analyzing regression errors in the NLP model updates.
We formulate the regression-free model updates into a constrained optimization problem.
We empirically analyze how model ensemble reduces regression.
arXiv Detail & Related papers (2021-05-07T03:33:00Z) - Positive-Congruent Training: Towards Regression-Free Model Updates [87.25247195148187]
In image classification, sample-wise inconsistencies appear as "negative flips"
A new model incorrectly predicts the output for a test sample that was correctly classified by the old (reference) model.
We propose a simple approach for PC training, Focal Distillation, which enforces congruence with the reference model.
arXiv Detail & Related papers (2020-11-18T09:00:44Z) - Towards Backward-Compatible Representation Learning [86.39292571306395]
We propose a way to learn visual features that are compatible with previously computed ones even when they have different dimensions.
This enables visual search systems to bypass computing new features for all previously seen images when updating the embedding models.
We propose a framework to train embedding models, called backward-compatible training (BCT), as a first step towards backward compatible representation learning.
arXiv Detail & Related papers (2020-03-26T14:34:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.