Mind the Backbone: Minimizing Backbone Distortion for Robust Object
Detection
- URL: http://arxiv.org/abs/2303.14744v2
- Date: Mon, 15 May 2023 18:52:42 GMT
- Title: Mind the Backbone: Minimizing Backbone Distortion for Robust Object
Detection
- Authors: Kuniaki Saito, Donghyun Kim, Piotr Teterwak, Rogerio Feris, Kate
Saenko
- Abstract summary: Building object detectors that are robust to domain shifts is critical for real-world applications.
We propose to use Relative Gradient Norm as a way to measure the vulnerability of a backbone to feature distortion.
We present recipes to boost OOD robustness for both types of backbones.
- Score: 52.355018626115346
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Building object detectors that are robust to domain shifts is critical for
real-world applications. Prior approaches fine-tune a pre-trained backbone and
risk overfitting it to in-distribution (ID) data and distorting features useful
for out-of-distribution (OOD) generalization. We propose to use Relative
Gradient Norm (RGN) as a way to measure the vulnerability of a backbone to
feature distortion, and show that high RGN is indeed correlated with lower OOD
performance. Our analysis of RGN yields interesting findings: some backbones
lose OOD robustness during fine-tuning, but others gain robustness because
their architecture prevents the parameters from changing too much from the
initial model. Given these findings, we present recipes to boost OOD robustness
for both types of backbones. Specifically, we investigate regularization and
architectural choices for minimizing gradient updates so as to prevent the
tuned backbone from losing generalizable features. Our proposed techniques
complement each other and show substantial improvements over baselines on
diverse architectures and datasets. Code is available at
https://github.com/VisionLearningGroup/mind_back.
Related papers
- Dissecting the Failure of Invariant Learning on Graphs [36.11431280689549]
We develop a Structural Causal Model (SCM) to theoretically dissect the performance of two prominent invariant learning methods.
We propose Cross-environment Intra-class Alignment (CIA), which explicitly eliminates spurious features by aligning cross-environment representations conditioned on the same class.
We further propose CIA-LRA (Localized Reweighting Alignment) that leverages the distribution of neighboring labels to selectively align node representations.
arXiv Detail & Related papers (2024-11-05T06:36:48Z) - Towards Robust Out-of-Distribution Generalization: Data Augmentation and Neural Architecture Search Approaches [4.577842191730992]
We study ways toward robust OoD generalization for deep learning.
We first propose a novel and effective approach to disentangle the spurious correlation between features that are not essential for recognition.
We then study the problem of strengthening neural architecture search in OoD scenarios.
arXiv Detail & Related papers (2024-10-25T20:50:32Z) - RPBG: Towards Robust Neural Point-based Graphics in the Wild [23.497693378880303]
We propose a new point-based neural re-rendering method, known as Point-based Graphics (RPBG)
RPBG stably outperforms the baseline by a large margin, and exhibits its great robustness over state-of-the-art NeRF-based variants.
arXiv Detail & Related papers (2024-05-09T10:35:44Z) - OOD-CV-v2: An extended Benchmark for Robustness to Out-of-Distribution
Shifts of Individual Nuisances in Natural Images [59.51657161097337]
OOD-CV-v2 is a benchmark dataset that includes out-of-distribution examples of 10 object categories in terms of pose, shape, texture, context and the weather conditions.
In addition to this novel dataset, we contribute extensive experiments using popular baseline methods.
arXiv Detail & Related papers (2023-04-17T20:39:25Z) - Energy-based Out-of-Distribution Detection for Graph Neural Networks [76.0242218180483]
We propose a simple, powerful and efficient OOD detection model for GNN-based learning on graphs, which we call GNNSafe.
GNNSafe achieves up to $17.0%$ AUROC improvement over state-of-the-arts and it could serve as simple yet strong baselines in such an under-developed area.
arXiv Detail & Related papers (2023-02-06T16:38:43Z) - Learning Target-aware Representation for Visual Tracking via Informative
Interactions [49.552877881662475]
We introduce a novel backbone architecture to improve target-perception ability of feature representation for tracking.
The proposed GIM module and InBN mechanism are general and applicable to different backbone types including CNN and Transformer.
arXiv Detail & Related papers (2022-01-07T16:22:27Z) - Robust lEarned Shrinkage-Thresholding (REST): Robust unrolling for
sparse recover [87.28082715343896]
We consider deep neural networks for solving inverse problems that are robust to forward model mis-specifications.
We design a new robust deep neural network architecture by applying algorithm unfolding techniques to a robust version of the underlying recovery problem.
The proposed REST network is shown to outperform state-of-the-art model-based and data-driven algorithms in both compressive sensing and radar imaging problems.
arXiv Detail & Related papers (2021-10-20T06:15:45Z) - Enhancing the Generalization for Intent Classification and Out-of-Domain
Detection in SLU [70.44344060176952]
Intent classification is a major task in spoken language understanding (SLU)
Recent works have shown that using extra data and labels can improve the OOD detection performance.
This paper proposes to train a model with only IND data while supporting both IND intent classification and OOD detection.
arXiv Detail & Related papers (2021-06-28T08:27:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.