Using Wavelets to Analyze Similarities in Image-Classification Datasets
- URL: http://arxiv.org/abs/2002.10257v2
- Date: Mon, 18 May 2020 01:42:54 GMT
- Title: Using Wavelets to Analyze Similarities in Image-Classification Datasets
- Authors: Roozbeh Yousefzadeh
- Abstract summary: We develop a practical and fast method to analyze the similarities in image classification datasets.
Our method uses wavelet decomposition of images and other numerical analysis tools, with no need for a pre-trained model.
We show that similar images in standard datasets can be identified in a few seconds, a significant speed-up compared to alternative methods in the literature.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning image classifiers usually rely on huge training sets and their
training process can be described as learning the similarities and differences
among training images. But, images in large training sets are not usually
studied from this perspective and fine-level similarities and differences among
images is usually overlooked. This is due to lack of fast and efficient
computational methods to analyze the contents of these datasets. Some studies
aim to identify the influential and redundant training images, but such methods
require a model that is already trained on the entire training set. Here, using
image processing and numerical analysis tools we develop a practical and fast
method to analyze the similarities in image classification datasets. We show
that such analysis can provide valuable insights about the datasets and the
classification task at hand, prior to training a model. Our method uses wavelet
decomposition of images and other numerical analysis tools, with no need for a
pre-trained model. Interestingly, the results we obtain corroborate the
previous results in the literature that analyzed the similarities using
pre-trained CNNs. We show that similar images in standard datasets (such as
CIFAR) can be identified in a few seconds, a significant speed-up compared to
alternative methods in the literature. By removing the computational speed
obstacle, it becomes practical to gain new insights about the contents of
datasets and the models trained on them. We show that similarities between
training and testing images may provide insights about the generalization of
models. Finally, we investigate the similarities between images in relation to
decision boundaries of a trained model.
Related papers
- Reinforcing Pre-trained Models Using Counterfactual Images [54.26310919385808]
This paper proposes a novel framework to reinforce classification models using language-guided generated counterfactual images.
We identify model weaknesses by testing the model using the counterfactual image dataset.
We employ the counterfactual images as an augmented dataset to fine-tune and reinforce the classification model.
arXiv Detail & Related papers (2024-06-19T08:07:14Z) - Scaling Laws of Synthetic Images for Model Training ... for Now [54.43596959598466]
We study the scaling laws of synthetic images generated by state of the art text-to-image models.
We observe that synthetic images demonstrate a scaling trend similar to, but slightly less effective than, real images in CLIP training.
arXiv Detail & Related papers (2023-12-07T18:59:59Z) - CorrEmbed: Evaluating Pre-trained Model Image Similarity Efficacy with a
Novel Metric [6.904776368895614]
We evaluate the viability of the image embeddings from pre-trained computer vision models using a novel approach named CorrEmbed.
Our approach computes the correlation between distances in image embeddings and distances in human-generated tag vectors.
Our method also identifies deviations from this pattern, providing insights into how different models capture high-level image features.
arXiv Detail & Related papers (2023-08-30T16:23:07Z) - Evaluating Data Attribution for Text-to-Image Models [62.844382063780365]
We evaluate attribution through "customization" methods, which tune an existing large-scale model toward a given exemplar object or style.
Our key insight is that this allows us to efficiently create synthetic images that are computationally influenced by the exemplar by construction.
By taking into account the inherent uncertainty of the problem, we can assign soft attribution scores over a set of training images.
arXiv Detail & Related papers (2023-06-15T17:59:51Z) - Compound Figure Separation of Biomedical Images: Mining Large Datasets
for Self-supervised Learning [12.445324044675116]
We introduce a simulation-based training framework that minimizes the need for resource extensive bounding box annotations.
We also propose a new side loss that is optimized for compound figure separation.
This is the first study that evaluates the efficacy of leveraging self-supervised learning with compound image separation.
arXiv Detail & Related papers (2022-08-30T16:02:34Z) - Meta Internal Learning [88.68276505511922]
Internal learning for single-image generation is a framework, where a generator is trained to produce novel images based on a single image.
We propose a meta-learning approach that enables training over a collection of images, in order to model the internal statistics of the sample image more effectively.
Our results show that the models obtained are as suitable as single-image GANs for many common image applications.
arXiv Detail & Related papers (2021-10-06T16:27:38Z) - Rectifying the Shortcut Learning of Background: Shared Object
Concentration for Few-Shot Image Recognition [101.59989523028264]
Few-Shot image classification aims to utilize pretrained knowledge learned from a large-scale dataset to tackle a series of downstream classification tasks.
We propose COSOC, a novel Few-Shot Learning framework, to automatically figure out foreground objects at both pretraining and evaluation stage.
arXiv Detail & Related papers (2021-07-16T07:46:41Z) - Evaluating Contrastive Models for Instance-based Image Retrieval [6.393147386784114]
We evaluate contrastive models for the task of image retrieval.
We find that models trained using contrastive methods perform on-par with (and outperforms) a pre-trained baseline trained on the ImageNet labels.
arXiv Detail & Related papers (2021-04-30T12:05:23Z) - Background Splitting: Finding Rare Classes in a Sea of Background [55.03789745276442]
We focus on the real-world problem of training accurate deep models for image classification of a small number of rare categories.
In these scenarios, almost all images belong to the background category in the dataset (>95% of the dataset is background)
We demonstrate that both standard fine-tuning approaches and state-of-the-art approaches for training on imbalanced datasets do not produce accurate deep models in the presence of this extreme imbalance.
arXiv Detail & Related papers (2020-08-28T23:05:15Z) - Multi-task pre-training of deep neural networks for digital pathology [8.74883469030132]
We first assemble and transform many digital pathology datasets into a pool of 22 classification tasks and almost 900k images.
We show that our models used as feature extractors either improve significantly over ImageNet pre-trained models or provide comparable performance.
arXiv Detail & Related papers (2020-05-05T08:50:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.