Artificial Perceptual Learning: Image Categorization with Weak
Supervision
- URL: http://arxiv.org/abs/2106.07559v1
- Date: Wed, 2 Jun 2021 21:12:20 GMT
- Title: Artificial Perceptual Learning: Image Categorization with Weak
Supervision
- Authors: Chengliang Tang, Mar\'ia Uriarte, Helen Jin, Douglas C. Morton, Tian
Zheng
- Abstract summary: We propose a novel machine learning framework, artificial learning (APL), to tackle the problem of weakly supervised image categorization.
APL is constructed using state-of-the-art machine learning algorithms as building blocks to mimic the cognitive development process known as infant categorization.
We validate the proposed framework using a small set of images with high quality human annotations and show that the proposed framework attains human-level cognitive economy.
- Score: 2.6774008509840996
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine learning has achieved much success on supervised learning tasks with
large sets of well-annotated training samples. However, in many practical
situations, such strong and high-quality supervision provided by training data
is unavailable due to the expensive and labor-intensive labeling process.
Automatically identifying and recognizing object categories in a large volume
of unlabeled images with weak supervision remains an important, yet unsolved
challenge in computer vision. In this paper, we propose a novel machine
learning framework, artificial perceptual learning (APL), to tackle the problem
of weakly supervised image categorization. The proposed APL framework is
constructed using state-of-the-art machine learning algorithms as building
blocks to mimic the cognitive development process known as infant
categorization. We develop and illustrate the proposed framework by
implementing a wide-field fine-grain ecological survey of tree species over an
8,000-hectare area of the El Yunque rainforest in Puerto Rico. It is based on
unlabeled high-resolution aerial images of the tree canopy. Misplaced
ground-based labels were available for less than 1% of these images, which
serve as the only weak supervision for this learning framework. We validate the
proposed framework using a small set of images with high quality human
annotations and show that the proposed framework attains human-level cognitive
economy.
Related papers
- RIGID: A Training-free and Model-Agnostic Framework for Robust AI-Generated Image Detection [60.960988614701414]
RIGID is a training-free and model-agnostic method for robust AI-generated image detection.
RIGID significantly outperforms existing trainingbased and training-free detectors.
arXiv Detail & Related papers (2024-05-30T14:49:54Z) - Cross-Level Multi-Instance Distillation for Self-Supervised Fine-Grained
Visual Categorization [41.86678318006878]
We propose a Cross-level Multi-instance Distillation (CMD) framework to tackle the challenge of fine-grained pre-text representation.
Our key idea is to consider the importance of each image patch in determining the fine-grained pre-text representation by multiple instance learning.
The proposed method outperforms the contemporary method by upto 10.14% and existing state-of-the-art self-supervised learning approaches by upto 19.78% on both top-1 accuracy and Rank-1 retrieval metric.
arXiv Detail & Related papers (2024-01-16T22:22:09Z) - Re-IQA: Unsupervised Learning for Image Quality Assessment in the Wild [38.197794061203055]
We propose a Mixture of Experts approach to train two separate encoders to learn high-level content and low-level image quality features in an unsupervised setting.
We deploy the complementary low and high-level image representations obtained from the Re-IQA framework to train a linear regression model.
Our method achieves state-of-the-art performance on multiple large-scale image quality assessment databases.
arXiv Detail & Related papers (2023-04-02T05:06:51Z) - Automated wildlife image classification: An active learning tool for
ecological applications [0.44970015278813025]
Wildlife camera trap images are being used extensively to investigate animal abundance, habitat associations, and behavior.
Artificial intelligence systems can take over this task but usually need a large number of already-labeled training images to achieve sufficient performance.
We propose a label-efficient learning strategy that enables researchers with small or medium-sized image databases to leverage the potential of modern machine learning.
arXiv Detail & Related papers (2023-03-28T08:51:15Z) - Weakly-supervised HOI Detection via Prior-guided Bi-level Representation
Learning [66.00600682711995]
Human object interaction (HOI) detection plays a crucial role in human-centric scene understanding and serves as a fundamental building-block for many vision tasks.
One generalizable and scalable strategy for HOI detection is to use weak supervision, learning from image-level annotations only.
This is inherently challenging due to ambiguous human-object associations, large search space of detecting HOIs and highly noisy training signal.
We develop a CLIP-guided HOI representation capable of incorporating the prior knowledge at both image level and HOI instance level, and adopt a self-taught mechanism to prune incorrect human-object associations.
arXiv Detail & Related papers (2023-03-02T14:41:31Z) - Urban feature analysis from aerial remote sensing imagery using
self-supervised and semi-supervised computer vision [8.124947412639704]
Analysis of overhead imagery using computer vision is a problem that has received considerable attention in academic literature.
These problems are addressed here through the development of a more generic framework, incorporating advances in representation learning.
The successful low-level detection of urban infrastructure evolution over a 10-year period from 60 million unlabeled images, exemplifies the substantial potential of our approach to advance quantitative urban research.
arXiv Detail & Related papers (2022-08-17T03:41:56Z) - Exploring CLIP for Assessing the Look and Feel of Images [87.97623543523858]
We introduce Contrastive Language-Image Pre-training (CLIP) models for assessing both the quality perception (look) and abstract perception (feel) of images in a zero-shot manner.
Our results show that CLIP captures meaningful priors that generalize well to different perceptual assessments.
arXiv Detail & Related papers (2022-07-25T17:58:16Z) - An Empirical Investigation of Representation Learning for Imitation [76.48784376425911]
Recent work in vision, reinforcement learning, and NLP has shown that auxiliary representation learning objectives can reduce the need for large amounts of expensive, task-specific data.
We propose a modular framework for constructing representation learning algorithms, then use our framework to evaluate the utility of representation learning for imitation.
arXiv Detail & Related papers (2022-05-16T11:23:42Z) - Region-level Active Learning for Cluttered Scenes [60.93811392293329]
We introduce a new strategy that subsumes previous Image-level and Object-level approaches into a generalized, Region-level approach.
We show that this approach significantly decreases labeling effort and improves rare object search on realistic data with inherent class-imbalance and cluttered scenes.
arXiv Detail & Related papers (2021-08-20T14:02:38Z) - Mixed Supervision Learning for Whole Slide Image Classification [88.31842052998319]
We propose a mixed supervision learning framework for super high-resolution images.
During the patch training stage, this framework can make use of coarse image-level labels to refine self-supervised learning.
A comprehensive strategy is proposed to suppress pixel-level false positives and false negatives.
arXiv Detail & Related papers (2021-07-02T09:46:06Z) - Unsupervised Image Classification for Deep Representation Learning [42.09716669386924]
We propose an unsupervised image classification framework without using embedding clustering.
Experiments on ImageNet dataset have been conducted to prove the effectiveness of our method.
arXiv Detail & Related papers (2020-06-20T02:57:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.