Reproduction of Lateral Inhibition-Inspired Convolutional Neural Network
for Visual Attention and Saliency Detection
- URL: http://arxiv.org/abs/2005.02184v1
- Date: Tue, 5 May 2020 13:55:47 GMT
- Title: Reproduction of Lateral Inhibition-Inspired Convolutional Neural Network
for Visual Attention and Saliency Detection
- Authors: Filip Marcinek
- Abstract summary: neural networks can be effectively confused with even natural images examples.
I suspect that the classification of an object is strongly influenced by the background pixels on which the object is located.
I analyze the above problem using for this purpose saliency maps created by the LICNN network.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, neural networks have continued to flourish, achieving high
efficiency in detecting relevant objects in photos or simply recognizing
(classifying) these objects - mainly using CNN networks. Current solutions,
however, are far from ideal, because it often turns out that network can be
effectively confused with even natural images examples. I suspect that the
classification of an object is strongly influenced by the background pixels on
which the object is located. In my work, I analyze the above problem using for
this purpose saliency maps created by the LICNN network. They are designed to
suppress the neurons surrounding the examined object and, consequently, reduce
the contribution of background pixels to the classifier predictions. My
experiments on the natural and adversarial images datasets show that, indeed,
there is a visible correlation between the background and the wrong-classified
foreground object. This behavior of the network is not supported by human
experience, because, for example, we do not confuse the yellow school bus with
the snow plow just because it is on the snowy background.
Related papers
- Visual Context-Aware Person Fall Detection [52.49277799455569]
We present a segmentation pipeline to semi-automatically separate individuals and objects in images.
Background objects such as beds, chairs, or wheelchairs can challenge fall detection systems, leading to false positive alarms.
We demonstrate that object-specific contextual transformations during training effectively mitigate this challenge.
arXiv Detail & Related papers (2024-04-11T19:06:36Z) - Understanding the Role of Pathways in a Deep Neural Network [4.456675543894722]
We analyze a convolutional neural network (CNN) trained in the classification task and present an algorithm to extract the diffusion pathways of individual pixels.
We find that the few largest pathways of an individual pixel from an image tend to cross the feature maps in each layer that is important for classification.
arXiv Detail & Related papers (2024-02-28T07:53:19Z) - Why do CNNs excel at feature extraction? A mathematical explanation [53.807657273043446]
We introduce a novel model for image classification, based on feature extraction, that can be used to generate images resembling real-world datasets.
In our proof, we construct piecewise linear functions that detect the presence of features, and show that they can be realized by a convolutional network.
arXiv Detail & Related papers (2023-07-03T10:41:34Z) - Background Invariant Classification on Infrared Imagery by Data
Efficient Training and Reducing Bias in CNNs [1.2891210250935146]
convolutional neural networks can classify objects in images very accurately.
It is well known that the attention of the network may not always be on the semantically important regions of the scene.
We propose a new two-step training procedure called textitsplit training to reduce this bias in CNNs on both Infrared imagery and RGB data.
arXiv Detail & Related papers (2022-01-22T23:29:42Z) - Learning to Detect Every Thing in an Open World [139.78830329914135]
We propose a simple yet surprisingly powerful data augmentation and training scheme we call Learning to Detect Every Thing (LDET)
To avoid suppressing hidden objects, background objects that are visible but unlabeled, we paste annotated objects on a background image sampled from a small region of the original image.
LDET leads to significant improvements on many datasets in the open world instance segmentation task.
arXiv Detail & Related papers (2021-12-03T03:56:06Z) - Anabranch Network for Camouflaged Object Segmentation [23.956327305907585]
This paper explores the camouflaged object segmentation problem, namely, segmenting the camouflaged object(s) for a given image.
To address this problem, we provide a new image dataset of camouflaged objects for benchmarking purposes.
In addition, we propose a general end-to-end network, called the Anabranch Network, that leverages both classification and segmentation tasks.
arXiv Detail & Related papers (2021-05-20T01:52:44Z) - Structure-Preserving Progressive Low-rank Image Completion for Defending
Adversarial Attacks [20.700098449823024]
Deep neural networks recognize objects by analyzing local image details and summarizing their information along the inference layers to derive the final decision.
Small sophisticated noise in the input images can accumulate along the network inference path and produce wrong decisions at the network output.
Human eyes recognize objects based on their global structure and semantic cues, instead of local image textures.
arXiv Detail & Related papers (2021-03-04T01:24:15Z) - Improving Object Detection in Art Images Using Only Style Transfer [5.156484100374058]
We propose and evaluate a process for training neural networks to localize objects - specifically people - in art images.
We generate a large dataset for training and validation by modifying the images in the COCO dataset using AdaIn style transfer.
The result is a significant improvement on the state of the art and a new way forward for creating datasets to train neural networks to process art images.
arXiv Detail & Related papers (2021-02-12T13:48:46Z) - Assessing The Importance Of Colours For CNNs In Object Recognition [70.70151719764021]
Convolutional neural networks (CNNs) have been shown to exhibit conflicting properties.
We demonstrate that CNNs often rely heavily on colour information while making a prediction.
We evaluate a model trained with congruent images on congruent, greyscale, and incongruent images.
arXiv Detail & Related papers (2020-12-12T22:55:06Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z) - WW-Nets: Dual Neural Networks for Object Detection [48.67090730174743]
We propose a new deep convolutional neural network framework that uses object location knowledge implicit in network connection weights to guide selective attention in object detection tasks.
Our approach is called What-Where Nets (WW-Nets), and it is inspired by the structure of human visual pathways.
arXiv Detail & Related papers (2020-05-15T21:16:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.