Robust Computer Vision in an Ever-Changing World: A Survey of Techniques
for Tackling Distribution Shifts
- URL: http://arxiv.org/abs/2312.01540v1
- Date: Sun, 3 Dec 2023 23:40:12 GMT
- Title: Robust Computer Vision in an Ever-Changing World: A Survey of Techniques
for Tackling Distribution Shifts
- Authors: Eashan Adhikarla, Kai Zhang, Jun Yu, Lichao Sun, John Nicholson and
Brian D. Davison
- Abstract summary: AI applications are becoming increasingly visible to the general public.
There is a notable gap between the theoretical assumptions researchers make about computer vision models and the reality those models face when deployed in the real world.
One of the critical reasons for this gap is a challenging problem known as distribution shift.
- Score: 20.17397328893533
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: AI applications are becoming increasingly visible to the general public.
There is a notable gap between the theoretical assumptions researchers make
about computer vision models and the reality those models face when deployed in
the real world. One of the critical reasons for this gap is a challenging
problem known as distribution shift. Distribution shifts tend to vary with
complexity of the data, dataset size, and application type. In our paper, we
discuss the identification of such a prominent gap, exploring the concept of
distribution shift and its critical significance. We provide an in-depth
overview of various types of distribution shifts, elucidate their distinctions,
and explore techniques within the realm of the data-centric domain employed to
address them. Distribution shifts can occur during every phase of the machine
learning pipeline, from the data collection stage to the stage of training a
machine learning model to the stage of final model deployment. As a result, it
raises concerns about the overall robustness of the machine learning techniques
for computer vision applications that are deployed publicly for consumers.
Different deep learning models each tailored for specific type of data and
tasks, architectural pipelines; highlighting how variations in data
preprocessing and feature extraction can impact robustness., data augmentation
strategies (e.g. geometric, synthetic and learning-based); demonstrating their
role in enhancing model generalization, and training mechanisms (e.g. transfer
learning, zero-shot) fall under the umbrella of data-centric methods. Each of
these components form an integral part of the neural-network we analyze
contributing uniquely to strengthening model robustness against distribution
shifts. We compare and contrast numerous AI models that are built for
mitigating shifts in hidden stratification and spurious correlations, ...
Related papers
- A Survey of Deep Graph Learning under Distribution Shifts: from Graph Out-of-Distribution Generalization to Adaptation [59.14165404728197]
We provide an up-to-date and forward-looking review of deep graph learning under distribution shifts.
Specifically, we cover three primary scenarios: graph OOD generalization, training-time graph OOD adaptation, and test-time graph OOD adaptation.
To provide a better understanding of the literature, we systematically categorize the existing models based on our proposed taxonomy.
arXiv Detail & Related papers (2024-10-25T02:39:56Z) - Learning Divergence Fields for Shift-Robust Graph Representations [73.11818515795761]
In this work, we propose a geometric diffusion model with learnable divergence fields for the challenging problem with interdependent data.
We derive a new learning objective through causal inference, which can guide the model to learn generalizable patterns of interdependence that are insensitive across domains.
arXiv Detail & Related papers (2024-06-07T14:29:21Z) - Machine Learning vs Deep Learning: The Generalization Problem [0.0]
This study investigates the comparative abilities of traditional machine learning (ML) models and deep learning (DL) algorithms in terms of extrapolation.
We present an empirical analysis where both ML and DL models are trained on an exponentially growing function and then tested on values outside the training domain.
Our findings suggest that deep learning models possess inherent capabilities to generalize beyond the training scope.
arXiv Detail & Related papers (2024-03-03T21:42:55Z) - Better, Not Just More: Data-Centric Machine Learning for Earth Observation [16.729827218159038]
We argue that a shift from a model-centric view to a complementary data-centric perspective is necessary for further improvements in accuracy, generalization ability, and real impact on end-user applications.
This work presents a definition as well as a precise categorization and overview of automated data-centric learning approaches for geospatial data.
arXiv Detail & Related papers (2023-12-08T19:24:05Z) - Fantastic Gains and Where to Find Them: On the Existence and Prospect of
General Knowledge Transfer between Any Pretrained Model [74.62272538148245]
We show that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other.
We investigate if it is possible to transfer such "complementary" knowledge from one model to another without performance degradation.
arXiv Detail & Related papers (2023-10-26T17:59:46Z) - ProtoVAE: Prototypical Networks for Unsupervised Disentanglement [1.6114012813668934]
We introduce a novel deep generative VAE-based model, ProtoVAE, that leverages a deep metric learning Prototypical network trained using self-supervision.
Our model is completely unsupervised and requires no priori knowledge of the dataset, including the number of factors.
We evaluate our proposed model on the benchmark dSprites, 3DShapes, and MPI3D disentanglement datasets.
arXiv Detail & Related papers (2023-05-16T01:29:26Z) - Synthetic-to-Real Domain Generalized Semantic Segmentation for 3D Indoor
Point Clouds [69.64240235315864]
This paper introduces the synthetic-to-real domain generalization setting to this task.
The domain gap between synthetic and real-world point cloud data mainly lies in the different layouts and point patterns.
Experiments on the synthetic-to-real benchmark demonstrate that both CINMix and multi-prototypes can narrow the distribution gap.
arXiv Detail & Related papers (2022-12-09T05:07:43Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - Deep invariant networks with differentiable augmentation layers [87.22033101185201]
Methods for learning data augmentation policies require held-out data and are based on bilevel optimization problems.
We show that our approach is easier and faster to train than modern automatic data augmentation techniques.
arXiv Detail & Related papers (2022-02-04T14:12:31Z) - Visual Representation Learning Does Not Generalize Strongly Within the
Same Domain [41.66817277929783]
We test whether 17 unsupervised, weakly supervised, and fully supervised representation learning approaches correctly infer the generative factors of variation in simple datasets.
We train and test 2000+ models and observe that all of them struggle to learn the underlying mechanism regardless of supervision signal and architectural bias.
arXiv Detail & Related papers (2021-07-17T11:24:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.