The MAMe Dataset: On the relevance of High Resolution and Variable Shape
image properties
- URL: http://arxiv.org/abs/2007.13693v3
- Date: Thu, 20 May 2021 10:57:06 GMT
- Title: The MAMe Dataset: On the relevance of High Resolution and Variable Shape
image properties
- Authors: Ferran Par\'es, Anna Arias-Duart, Dario Garcia-Gasulla, Gema
Campo-Franc\'es, Nina Viladrich, Eduard Ayguad\'e, Jes\'us Labarta
- Abstract summary: We introduce the MAMe dataset, an image classification dataset with remarkable high resolution and variable shape properties.
The MAMe dataset contains thousands of artworks from three different museums.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In the image classification task, the most common approach is to resize all
images in a dataset to a unique shape, while reducing their precision to a size
which facilitates experimentation at scale. This practice has benefits from a
computational perspective, but it entails negative side-effects on performance
due to loss of information and image deformation. In this work we introduce the
MAMe dataset, an image classification dataset with remarkable high resolution
and variable shape properties. The goal of MAMe is to provide a tool for
studying the impact of such properties in image classification, while
motivating research in the field. The MAMe dataset contains thousands of
artworks from three different museums, and proposes a classification task
consisting on differentiating between 29 mediums (i.e. materials and
techniques) supervised by art experts. After reviewing the singularity of MAMe
in the context of current image classification tasks, a thorough description of
the task is provided, together with dataset statistics. Experiments are
conducted to evaluate the impact of using high resolution images, variable
shape inputs and both properties at the same time. Results illustrate the
positive impact in performance when using high resolution images, while
highlighting the lack of solutions to exploit variable shapes. An additional
experiment exposes the distinctiveness between the MAMe dataset and the
prototypical ImageNet dataset. Finally, the baselines are inspected using
explainability methods and expert knowledge, to gain insights on the challenges
that remain ahead.
Related papers
- Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - SeeBel: Seeing is Believing [0.9790236766474201]
We propose three visualizations that enable users to compare dataset statistics and AI performance for segmenting all images.
Our project tries to further increase the interpretability of the trained AI model for segmentation by visualizing its image attention weights.
We propose to conduct surveys on real users to study the efficacy of our visualization tool in computer vision and AI domain.
arXiv Detail & Related papers (2023-12-18T05:11:00Z) - Improving Human-Object Interaction Detection via Virtual Image Learning [68.56682347374422]
Human-Object Interaction (HOI) detection aims to understand the interactions between humans and objects.
In this paper, we propose to alleviate the impact of such an unbalanced distribution via Virtual Image Leaning (VIL)
A novel label-to-image approach, Multiple Steps Image Creation (MUSIC), is proposed to create a high-quality dataset that has a consistent distribution with real images.
arXiv Detail & Related papers (2023-08-04T10:28:48Z) - High-Quality Entity Segmentation [110.55724145851725]
CropFormer is designed to tackle the intractability of instance-level segmentation on high-resolution images.
It improves mask prediction by fusing high-res image crops that provide more fine-grained image details and the full image.
With CropFormer, we achieve a significant AP gain of $1.9$ on the challenging entity segmentation task.
arXiv Detail & Related papers (2022-11-10T18:58:22Z) - Object-aware Contrastive Learning for Debiased Scene Representation [74.30741492814327]
We develop a novel object-aware contrastive learning framework that localizes objects in a self-supervised manner.
We also introduce two data augmentations based on ContraCAM, object-aware random crop and background mixup, which reduce contextual and background biases during contrastive self-supervised learning.
arXiv Detail & Related papers (2021-07-30T19:24:07Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Factors of Influence for Transfer Learning across Diverse Appearance
Domains and Task Types [50.1843146606122]
A simple form of transfer learning is common in current state-of-the-art computer vision models.
Previous systematic studies of transfer learning have been limited and the circumstances in which it is expected to work are not fully understood.
In this paper we carry out an extensive experimental exploration of transfer learning across vastly different image domains.
arXiv Detail & Related papers (2021-03-24T16:24:20Z) - Image Augmentation for Multitask Few-Shot Learning: Agricultural Domain
Use-Case [0.0]
This paper challenges small and imbalanced datasets based on the example of a plant phenomics domain.
We introduce an image augmentation framework, which enables us to extremely enlarge the number of training samples.
We prove that our augmentation method increases model performance when only a few training samples are available.
arXiv Detail & Related papers (2021-02-24T14:08:34Z) - Multi-scale Information Assembly for Image Matting [35.43994064645042]
We propose a multi-scale information assembly framework (MSIA-matte) to pull out high-quality alpha mattes from single RGB images.
We can achieve state-of-the-art performance compared to most existing matting networks.
arXiv Detail & Related papers (2021-01-07T06:15:48Z) - Unsupervised machine learning via transfer learning and k-means
clustering to classify materials image data [0.0]
This paper demonstrates how to construct, use, and evaluate a high performance unsupervised machine learning system for classifying images.
We use the VGG16 convolutional neural network pre-trained on the ImageNet dataset of natural images to extract feature representations for each micrograph.
The approach achieves $99.4% pm 0.16%$ accuracy, and the resulting model can be used to classify new images without retraining.
arXiv Detail & Related papers (2020-07-16T14:36:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.