Few-Cost Salient Object Detection with Adversarial-Paced Learning
- URL: http://arxiv.org/abs/2104.01928v1
- Date: Mon, 5 Apr 2021 14:15:49 GMT
- Title: Few-Cost Salient Object Detection with Adversarial-Paced Learning
- Authors: Dingwen Zhang, Haibin Tian, and Jungong Han
- Abstract summary: This paper proposes to learn the effective salient object detection model based on the manual annotation on a few training images only.
We name this task as the few-cost salient object detection and propose an adversarial-paced learning (APL)-based framework to facilitate the few-cost learning scenario.
- Score: 95.0220555274653
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Detecting and segmenting salient objects from given image scenes has received
great attention in recent years. A fundamental challenge in training the
existing deep saliency detection models is the requirement of large amounts of
annotated data. While gathering large quantities of training data becomes cheap
and easy, annotating the data is an expensive process in terms of time, labor
and human expertise. To address this problem, this paper proposes to learn the
effective salient object detection model based on the manual annotation on a
few training images only, thus dramatically alleviating human labor in training
models. To this end, we name this task as the few-cost salient object detection
and propose an adversarial-paced learning (APL)-based framework to facilitate
the few-cost learning scenario. Essentially, APL is derived from the self-paced
learning (SPL) regime but it infers the robust learning pace through the
data-driven adversarial learning mechanism rather than the heuristic design of
the learning regularizer. Comprehensive experiments on four widely-used
benchmark datasets demonstrate that the proposed method can effectively
approach to the existing supervised deep salient object detection models with
only 1k human-annotated training images. The project page is available at
https://github.com/hb-stone/FC-SOD.
Related papers
- EfficientTrain++: Generalized Curriculum Learning for Efficient Visual Backbone Training [79.96741042766524]
We reformulate the training curriculum as a soft-selection function.
We show that exposing the contents of natural images can be readily achieved by the intensity of data augmentation.
The resulting method, EfficientTrain++, is simple, general, yet surprisingly effective.
arXiv Detail & Related papers (2024-05-14T17:00:43Z) - Dynamic Task and Weight Prioritization Curriculum Learning for
Multimodal Imagery [0.5439020425819]
This paper explores post-disaster analytics using multimodal deep learning models trained with curriculum learning method.
Curriculum learning emulates the progressive learning sequence in human education by training deep learning models on increasingly complex data.
arXiv Detail & Related papers (2023-10-29T18:46:33Z) - Improved Region Proposal Network for Enhanced Few-Shot Object Detection [23.871860648919593]
Few-shot object detection (FSOD) methods have emerged as a solution to the limitations of classic object detection approaches.
We develop a semi-supervised algorithm to detect and then utilize unlabeled novel objects as positive samples during the FSOD training stage.
Our improved hierarchical sampling strategy for the region proposal network (RPN) also boosts the perception of the object detection model for large objects.
arXiv Detail & Related papers (2023-08-15T02:35:59Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - EfficientTrain: Exploring Generalized Curriculum Learning for Training
Visual Backbones [80.662250618795]
This paper presents a new curriculum learning approach for the efficient training of visual backbones (e.g., vision Transformers)
As an off-the-shelf method, it reduces the wall-time training cost of a wide variety of popular models by >1.5x on ImageNet-1K/22K without sacrificing accuracy.
arXiv Detail & Related papers (2022-11-17T17:38:55Z) - Self-Supervised Learning of Multi-Object Keypoints for Robotic
Manipulation [8.939008609565368]
In this paper, we demonstrate the efficacy of learning image keypoints via the Dense Correspondence pretext task for downstream policy learning.
We evaluate our approach on diverse robot manipulation tasks, compare it to other visual representation learning approaches, and demonstrate its flexibility and effectiveness for sample-efficient policy learning.
arXiv Detail & Related papers (2022-05-17T13:15:07Z) - Learning from Few Examples: A Summary of Approaches to Few-Shot Learning [3.6930948691311016]
Few-Shot Learning refers to the problem of learning the underlying pattern in the data just from a few training samples.
Deep learning solutions suffer from data hunger and extensively high computation time and resources.
Few-shot learning that could drastically reduce the turnaround time of building machine learning applications emerges as a low-cost solution.
arXiv Detail & Related papers (2022-03-07T23:15:21Z) - What Makes Good Contrastive Learning on Small-Scale Wearable-based
Tasks? [59.51457877578138]
We study contrastive learning on the wearable-based activity recognition task.
This paper presents an open-source PyTorch library textttCL-HAR, which can serve as a practical tool for researchers.
arXiv Detail & Related papers (2022-02-12T06:10:15Z) - Puzzle-AE: Novelty Detection in Images through Solving Puzzles [8.999416735254586]
U-Net is proved to be effective for this purpose but overfits on the training data if trained by just using reconstruction error similar to other AE-based frameworks.
We show that training U-Nets based on this task is an effective remedy that prevents overfitting and facilitates learning beyond pixel-level features.
We propose adversarial robust training as an effective automatic shortcut removal.
arXiv Detail & Related papers (2020-08-29T10:53:55Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.