Joint Salient Object Detection and Camouflaged Object Detection via
Uncertainty-aware Learning
- URL: http://arxiv.org/abs/2307.04651v1
- Date: Mon, 10 Jul 2023 15:49:37 GMT
- Title: Joint Salient Object Detection and Camouflaged Object Detection via
Uncertainty-aware Learning
- Authors: Aixuan Li, Jing Zhang, Yunqiu Lv, Tong Zhang, Yiran Zhong, Mingyi He,
Yuchao Dai
- Abstract summary: We introduce an uncertainty-aware learning pipeline to explore the contradictory information of salient object detection (SOD) and camouflaged object detection (COD)
Our solution leads to both state-of-the-art performance and informative uncertainty estimation.
- Score: 47.253370009231645
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Salient objects attract human attention and usually stand out clearly from
their surroundings. In contrast, camouflaged objects share similar colors or
textures with the environment. In this case, salient objects are typically
non-camouflaged, and camouflaged objects are usually not salient. Due to this
inherent contradictory attribute, we introduce an uncertainty-aware learning
pipeline to extensively explore the contradictory information of salient object
detection (SOD) and camouflaged object detection (COD) via data-level and
task-wise contradiction modeling. We first exploit the dataset correlation of
these two tasks and claim that the easy samples in the COD dataset can serve as
hard samples for SOD to improve the robustness of the SOD model. Based on the
assumption that these two models should lead to activation maps highlighting
different regions of the same input image, we further introduce a contrastive
module with a joint-task contrastive learning framework to explicitly model the
contradictory attributes of these two tasks. Different from conventional
intra-task contrastive learning for unsupervised representation learning, our
contrastive module is designed to model the task-wise correlation, leading to
cross-task representation learning. To better understand the two tasks from the
perspective of uncertainty, we extensively investigate the uncertainty
estimation techniques for modeling the main uncertainties of the two tasks,
namely task uncertainty (for SOD) and data uncertainty (for COD), and aiming to
effectively estimate the challenging regions for each task to achieve
difficulty-aware learning. Experimental results on benchmark datasets
demonstrate that our solution leads to both state-of-the-art performance and
informative uncertainty estimation.
Related papers
- Uncertainty-Guided Appearance-Motion Association Network for Out-of-Distribution Action Detection [4.938957922033169]
Out-of-distribution (OOD) detection targets to detect and reject test samples with semantic shifts.
We propose a novel Uncertainty-Guided Appearance-Motion Association Network (UAAN)
We show that UAAN beats state-of-the-art methods by a significant margin, illustrating its effectiveness.
arXiv Detail & Related papers (2024-09-16T02:53:49Z) - DeTra: A Unified Model for Object Detection and Trajectory Forecasting [68.85128937305697]
Our approach formulates the union of the two tasks as a trajectory refinement problem.
To tackle this unified task, we design a refinement transformer that infers the presence, pose, and multi-modal future behaviors of objects.
In our experiments, we observe that ourmodel outperforms the state-of-the-art on Argoverse 2 Sensor and Open dataset.
arXiv Detail & Related papers (2024-06-06T18:12:04Z) - Few-shot Oriented Object Detection with Memorable Contrastive Learning in Remote Sensing Images [11.217630579076237]
Few-shot object detection (FSOD) has garnered significant research attention in the field of remote sensing.
We propose a novel FSOD method for remote sensing images called Few-shot Oriented object detection with Memorable Contrastive learning (FOMC)
Specifically, we employ oriented bounding boxes instead of traditional horizontal bounding boxes to learn a better feature representation for arbitrary-oriented aerial objects.
arXiv Detail & Related papers (2024-03-20T08:15:18Z) - Practical Edge Detection via Robust Collaborative Learning [11.176517889212015]
Edge detection is a core component in a wide range of visionoriented tasks.
To achieve the goal, two key issues should be concerned.
How to mitigate deep edge models from inefficient pre-trained backbones.
How to liberate the negative influence from noisy or even wrong labels in training data.
arXiv Detail & Related papers (2023-08-27T12:12:27Z) - Suspected Object Matters: Rethinking Model's Prediction for One-stage
Visual Grounding [93.82542533426766]
We propose a Suspected Object Transformation mechanism (SOT) to encourage the target object selection among the suspected ones.
SOT can be seamlessly integrated into existing CNN and Transformer-based one-stage visual grounders.
Extensive experiments demonstrate the effectiveness of our proposed method.
arXiv Detail & Related papers (2022-03-10T06:41:07Z) - Object-aware Contrastive Learning for Debiased Scene Representation [74.30741492814327]
We develop a novel object-aware contrastive learning framework that localizes objects in a self-supervised manner.
We also introduce two data augmentations based on ContraCAM, object-aware random crop and background mixup, which reduce contextual and background biases during contrastive self-supervised learning.
arXiv Detail & Related papers (2021-07-30T19:24:07Z) - Uncertainty-aware Joint Salient Object and Camouflaged Object Detection [43.01556978979627]
We propose a paradigm of leveraging the contradictory information to enhance the detection ability of both salient object detection and camouflaged object detection.
We introduce a similarity measure module to explicitly model the contradicting attributes of these two tasks.
Considering the uncertainty of labeling in both tasks' datasets, we propose an adversarial learning network to achieve both higher order similarity measure and network confidence estimation.
arXiv Detail & Related papers (2021-04-06T16:05:10Z) - Reinforcement Learning for Sparse-Reward Object-Interaction Tasks in a
First-person Simulated 3D Environment [73.9469267445146]
First-person object-interaction tasks in high-fidelity, 3D, simulated environments such as the AI2Thor pose significant sample-efficiency challenges for reinforcement learning agents.
We show that one can learn object-interaction tasks from scratch without supervision by learning an attentive object-model as an auxiliary task.
arXiv Detail & Related papers (2020-10-28T19:27:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.