Now You See Me: Robust approach to Partial Occlusions
- URL: http://arxiv.org/abs/2304.11779v2
- Date: Tue, 25 Apr 2023 11:45:50 GMT
- Title: Now You See Me: Robust approach to Partial Occlusions
- Authors: Karthick Prasad Gunasekaran, Nikita Jaiman
- Abstract summary: Occlusions of objects is one of the indispensable problems in Computer vision.
This paper introduces our own synthetically created dataset by utilising Stanford Car dataset.
We conduct a comprehensive analysis using various state of the art CNN models such as VGG-19, ResNet 50/101, GoogleNet, DenseNet 121.
- Score: 0.15229257192293202
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Occlusions of objects is one of the indispensable problems in Computer
vision. While Convolutional Neural Net-works (CNNs) provide various state of
the art approaches for regular image classification, they however, prove to be
not as effective for the classification of images with partial occlusions.
Partial occlusion is scenario where an object is occluded partially by some
other object/space. This problem when solved,holds tremendous potential to
facilitate various scenarios. We in particular are interested in autonomous
driving scenario and its implications in the same. Autonomous vehicle research
is one of the hot topics of this decade, there are ample situations of partial
occlusions of a driving sign or a person or other objects at different angles.
Considering its prime importance in situations which can be further extended to
video analytics of traffic data to handle crimes, anticipate income levels of
various groups etc.,this holds the potential to be exploited in many ways. In
this paper, we introduce our own synthetically created dataset by utilising
Stanford Car Dataset and adding occlusions of various sizes and nature to it.
On this created dataset, we conducted a comprehensive analysis using various
state of the art CNN models such as VGG-19, ResNet 50/101, GoogleNet, DenseNet
121. We further in depth study the effect of varying occlusion proportions and
nature on the performance of these models by fine tuning and training these
from scratch on dataset and how is it likely to perform when trained in
different scenarios, i.e., performance when training with occluded images and
unoccluded images, which model is more robust to partial occlusions and soon.
Related papers
- Are Deep Learning Models Robust to Partial Object Occlusion in Visual Recognition Tasks? [4.9260675787714]
Image classification models, including convolutional neural networks (CNNs), perform well on a variety of classification tasks but struggle under partial occlusion.
We contribute the Image Recognition Under Occlusion (IRUO) dataset, based on the recently developed Occluded Video Instance (IRUO) dataset (arXiv:2102.01558)
We find that modern CNN-based models show improved recognition accuracy on occluded images compared to earlier CNN-based models, and ViT-based models are more accurate than CNN-based models on occluded images.
arXiv Detail & Related papers (2024-09-16T23:21:22Z) - Visual Context-Aware Person Fall Detection [52.49277799455569]
We present a segmentation pipeline to semi-automatically separate individuals and objects in images.
Background objects such as beds, chairs, or wheelchairs can challenge fall detection systems, leading to false positive alarms.
We demonstrate that object-specific contextual transformations during training effectively mitigate this challenge.
arXiv Detail & Related papers (2024-04-11T19:06:36Z) - Identifying and Mitigating Model Failures through Few-shot CLIP-aided
Diffusion Generation [65.268245109828]
We propose an end-to-end framework to generate text descriptions of failure modes associated with spurious correlations.
These descriptions can be used to generate synthetic data using generative models, such as diffusion models.
Our experiments have shown remarkable textbfimprovements in accuracy ($sim textbf21%$) on hard sub-populations.
arXiv Detail & Related papers (2023-12-09T04:43:49Z) - Finding Differences Between Transformers and ConvNets Using
Counterfactual Simulation Testing [82.67716657524251]
We present a counterfactual framework that allows us to study the robustness of neural networks with respect to naturalistic variations.
Our method allows for a fair comparison of the robustness of recently released, state-of-the-art Convolutional Neural Networks and Vision Transformers.
arXiv Detail & Related papers (2022-11-29T18:59:23Z) - CARLA-GeAR: a Dataset Generator for a Systematic Evaluation of
Adversarial Robustness of Vision Models [61.68061613161187]
This paper presents CARLA-GeAR, a tool for the automatic generation of synthetic datasets for evaluating the robustness of neural models against physical adversarial patches.
The tool is built on the CARLA simulator, using its Python API, and allows the generation of datasets for several vision tasks in the context of autonomous driving.
The paper presents an experimental study to evaluate the performance of some defense methods against such attacks, showing how the datasets generated with CARLA-GeAR might be used in future work as a benchmark for adversarial defense in the real world.
arXiv Detail & Related papers (2022-06-09T09:17:38Z) - Video Class Agnostic Segmentation with Contrastive Learning for
Autonomous Driving [13.312978643938202]
We propose a novel auxiliary contrastive loss to learn the segmentation of known classes and unknown objects.
Unlike previous work in contrastive learning that samples the anchor, positive and negative examples on an image level, our contrastive learning method leverages pixel-wise semantic and temporal guidance.
We release a large-scale synthetic dataset for different autonomous driving scenarios that includes distinct and rare unknown objects.
arXiv Detail & Related papers (2021-05-07T23:07:06Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Contemplating real-world object classification [53.10151901863263]
We reanalyze the ObjectNet dataset recently proposed by Barbu et al. containing objects in daily life situations.
We find that applying deep models to the isolated objects, rather than the entire scene as is done in the original paper, results in around 20-30% performance improvement.
arXiv Detail & Related papers (2021-03-08T23:29:59Z) - Robust Instance Segmentation through Reasoning about Multi-Object
Occlusion [9.536947328412198]
We propose a deep network for multi-object instance segmentation that is robust to occlusion.
Our work builds on Compositional Networks, which learn a generative model of neural feature activations to locate occluders.
In particular, we obtain feed-forward predictions of the object classes and their instance and occluder segmentations.
arXiv Detail & Related papers (2020-12-03T17:41:55Z) - Multimodal Prototypical Networks for Few-shot Learning [20.100480009813953]
Cross-modal feature generation framework is used to enrich the low populated embedding space in few-shot scenarios.
We show that in such cases nearest neighbor classification is a viable approach and outperform state-of-the-art single-modal and multimodal few-shot learning methods.
arXiv Detail & Related papers (2020-11-17T19:32:59Z) - Compositional Convolutional Neural Networks: A Robust and Interpretable
Model for Object Recognition under Occlusion [21.737411464598797]
We show that black-box deep convolutional neural networks (DCNNs) have only limited robustness to partial occlusion.
We overcome these limitations by unifying DCNNs with part-based models into Compositional Convolutional Neural Networks (CompositionalNets)
Our experiments show that CompositionalNets improve by a large margin over their non-compositional counterparts at classifying and detecting partially occluded objects.
arXiv Detail & Related papers (2020-06-28T08:18:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.