Minimum Description Length Clustering to Measure Meaningful Image
Complexity
- URL: http://arxiv.org/abs/2306.14937v3
- Date: Sat, 19 Aug 2023 07:48:41 GMT
- Title: Minimum Description Length Clustering to Measure Meaningful Image
Complexity
- Authors: Louis Mahon, Thomas Lukasiewicz
- Abstract summary: We present a new image complexity metric through hierarchical clustering of patches.
We use the minimum description length principle to determine the number of clusters and designate certain points as outliers and, hence, correctly assign white noise a low score.
- Score: 65.268245109828
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing image complexity metrics cannot distinguish meaningful content from
noise. This means that white noise images, which contain no meaningful
information, are judged as highly complex. We present a new image complexity
metric through hierarchical clustering of patches. We use the minimum
description length principle to determine the number of clusters and designate
certain points as outliers and, hence, correctly assign white noise a low
score. The presented method has similarities to theoretical ideas for measuring
meaningful complexity. We conduct experiments on seven different sets of
images, which show that our method assigns the most accurate scores to all
images considered. Additionally, comparing the different levels of the
hierarchy of clusters can reveal how complexity manifests at different scales,
from local detail to global structure. We then present ablation studies showing
the contribution of the components of our method, and that it continues to
assign reasonable scores when the inputs are modified in certain ways,
including the addition of Gaussian noise and the lowering of the resolution.
Related papers
- Contrastive Mean-Shift Learning for Generalized Category Discovery [45.19923199324919]
We address the problem of generalized category discovery (GCD)
We revisit the mean-shift algorithm, i.e., a powerful technique for mode seeking, and incorporate it into a contrastive learning framework.
The proposed method, dubbed Contrastive Mean-Shift (CMS) learning, trains an image encoder to produce representations with better clustering properties.
arXiv Detail & Related papers (2024-04-15T04:31:24Z) - Multi-Granularity Denoising and Bidirectional Alignment for Weakly
Supervised Semantic Segmentation [75.32213865436442]
We propose an end-to-end multi-granularity denoising and bidirectional alignment (MDBA) model to alleviate the noisy label and multi-class generalization issues.
The MDBA model can reach the mIoU of 69.5% and 70.2% on validation and test sets for the PASCAL VOC 2012 dataset.
arXiv Detail & Related papers (2023-05-09T03:33:43Z) - Learning to Annotate Part Segmentation with Gradient Matching [58.100715754135685]
This paper focuses on tackling semi-supervised part segmentation tasks by generating high-quality images with a pre-trained GAN.
In particular, we formulate the annotator learning as a learning-to-learn problem.
We show that our method can learn annotators from a broad range of labelled images including real images, generated images, and even analytically rendered images.
arXiv Detail & Related papers (2022-11-06T01:29:22Z) - MD-Net: Multi-Detector for Local Feature Extraction [0.0]
We propose a deep feature extraction network capable of detecting a predefined number of complementary sets of keypoints at each image.
We train our network to predict the keypoints and compute the corresponding descriptors jointly.
With extensive experiments we show that our network, trained only on synthetically warped images, achieves competitive results on 3D reconstruction and re-localization tasks.
arXiv Detail & Related papers (2022-08-10T13:52:31Z) - Deep Semantic Statistics Matching (D2SM) Denoising Network [70.01091467628068]
We introduce the Deep Semantic Statistics Matching (D2SM) Denoising Network.
It exploits semantic features of pretrained classification networks, then it implicitly matches the probabilistic distribution of clear images at the semantic feature space.
By learning to preserve the semantic distribution of denoised images, we empirically find our method significantly improves the denoising capabilities of networks.
arXiv Detail & Related papers (2022-07-19T14:35:42Z) - Exploring Set Similarity for Dense Self-supervised Representation
Learning [96.35286140203407]
We propose to explore textbfset textbfsimilarity (SetSim) for dense self-supervised representation learning.
We generalize pixel-wise similarity learning to set-wise one to improve the robustness because sets contain more semantic and structure information.
Specifically, by resorting to attentional features of views, we establish corresponding sets, thus filtering out noisy backgrounds that may cause incorrect correspondences.
arXiv Detail & Related papers (2021-07-19T09:38:27Z) - Grafit: Learning fine-grained image representations with coarse labels [114.17782143848315]
This paper tackles the problem of learning a finer representation than the one provided by training labels.
By jointly leveraging the coarse labels and the underlying fine-grained latent space, it significantly improves the accuracy of category-level retrieval methods.
arXiv Detail & Related papers (2020-11-25T19:06:26Z) - Realizing Pixel-Level Semantic Learning in Complex Driving Scenes based
on Only One Annotated Pixel per Class [17.481116352112682]
We propose a new semantic segmentation task under complex driving scenes based on weakly supervised condition.
A three step process is built for pseudo labels generation, which progressively implement optimal feature representation for each category.
Experiments on Cityscapes dataset demonstrate that the proposed method provides a feasible way to solve weakly supervised semantic segmentation task.
arXiv Detail & Related papers (2020-03-10T12:57:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.