Building Flyweight FLIM-based CNNs with Adaptive Decoding for Object
Detection
- URL: http://arxiv.org/abs/2306.14840v2
- Date: Thu, 5 Oct 2023 09:22:34 GMT
- Title: Building Flyweight FLIM-based CNNs with Adaptive Decoding for Object
Detection
- Authors: Leonardo de Melo Joao, Azael de Melo e Sousa, Bianca Martins dos
Santos, Silvio Jamil Ferzoli Guimaraes, Jancarlo Ferreira Gomes, Ewa Kijak,
Alexandre Xavier Falcao
- Abstract summary: This work presents a method to build a Convolutional Neural Network (CNN) layer by layer for object detection from user-drawn markers.
We address the detection of Schistosomiasis mansoni eggs in microscopy images of fecal samples, and the detection of ships in satellite images.
Our CNN weighs thousands of times less than SOTA object detectors, being suitable for CPU execution and showing superior or equivalent performance to three methods in five measures.
- Score: 40.97322222472642
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: State-of-the-art (SOTA) object detection methods have succeeded in several
applications at the price of relying on heavyweight neural networks, which
makes them inefficient and inviable for many applications with computational
resource constraints. This work presents a method to build a Convolutional
Neural Network (CNN) layer by layer for object detection from user-drawn
markers on discriminative regions of representative images. We address the
detection of Schistosomiasis mansoni eggs in microscopy images of fecal
samples, and the detection of ships in satellite images as application
examples. We could create a flyweight CNN without backpropagation from very few
input images. Our method explores a recent methodology, Feature Learning from
Image Markers (FLIM), to build convolutional feature extractors (encoders) from
marker pixels. We extend FLIM to include a single-layer adaptive decoder, whose
weights vary with the input image -- a concept never explored in CNNs. Our CNN
weighs thousands of times less than SOTA object detectors, being suitable for
CPU execution and showing superior or equivalent performance to three methods
in five measures.
Related papers
- Deep Dynamic Scene Deblurring from Optical Flow [53.625999196063574]
Deblurring can provide visually more pleasant pictures and make photography more convenient.
It is difficult to model the non-uniform blur mathematically.
We develop a convolutional neural network (CNN) to restore the sharp images from the deblurred features.
arXiv Detail & Related papers (2023-01-18T06:37:21Z) - Paint and Distill: Boosting 3D Object Detection with Semantic Passing
Network [70.53093934205057]
3D object detection task from lidar or camera sensors is essential for autonomous driving.
We propose a novel semantic passing framework, named SPNet, to boost the performance of existing lidar-based 3D detection models.
arXiv Detail & Related papers (2022-07-12T12:35:34Z) - New SAR target recognition based on YOLO and very deep multi-canonical
correlation analysis [0.1503974529275767]
This paper proposes a robust feature extraction method for SAR image target classification by adaptively fusing effective features from different CNN layers.
Experiments on the MSTAR dataset demonstrate that the proposed method outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2021-10-28T18:10:26Z) - Issues in Object Detection in Videos using Common Single-Image CNNs [0.0]
Object detection is used in many applications such as industrial process, medical imaging analysis, and autonomous vehicles.
For applications such as autonomous vehicles, it is crucial that the object detection system can identify objects through multiple frames in video.
There are many neural networks that have been used for object detection and if there was a way of connecting objects between frames then these problems could be eliminated.
A dataset must be created with images that represent consecutive video frames and have matching ground-truth layers.
arXiv Detail & Related papers (2021-05-26T20:33:51Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Learning Hybrid Representations for Automatic 3D Vessel Centerline
Extraction [57.74609918453932]
Automatic blood vessel extraction from 3D medical images is crucial for vascular disease diagnoses.
Existing methods may suffer from discontinuities of extracted vessels when segmenting such thin tubular structures from 3D images.
We argue that preserving the continuity of extracted vessels requires to take into account the global geometry.
We propose a hybrid representation learning approach to address this challenge.
arXiv Detail & Related papers (2020-12-14T05:22:49Z) - Multi-pooled Inception features for no-reference image quality
assessment [0.0]
We propose a new approach for image quality assessment using convolutional neural networks (CNNs)
In contrast to previous methods, we do not take patches from the input image. Instead, the input image is treated as a whole and is run through a pretrained CNN body to extract resolution-independent, multi-level deep features.
We demonstrate that our best proposal - called MultiGAP-NRIQA - is able to provide state-of-the-art results on three benchmark IQA databases.
arXiv Detail & Related papers (2020-11-10T15:09:49Z) - Multiscale Detection of Cancerous Tissue in High Resolution Slide Scans [0.0]
We present an algorithm for multi-scale tumor (chimeric cell) detection in high resolution slide scans.
Our approach modifies the effective receptive field at different layers in a CNN so that objects with a broad range of varying scales can be detected in a single forward pass.
arXiv Detail & Related papers (2020-10-01T18:56:46Z) - Learning CNN filters from user-drawn image markers for coconut-tree
image classification [78.42152902652215]
We present a method that needs a minimal set of user-selected images to train the CNN's feature extractor.
The method learns the filters of each convolutional layer from user-drawn markers in image regions that discriminate classes.
It does not rely on optimization based on backpropagation, and we demonstrate its advantages on the binary classification of coconut-tree aerial images.
arXiv Detail & Related papers (2020-08-08T15:50:23Z) - Verification of Deep Convolutional Neural Networks Using ImageStars [10.44732293654293]
Convolutional Neural Networks (CNN) have redefined the state-of-the-art in many real-world applications.
CNNs are vulnerable to adversarial attacks, where slight changes to their inputs may lead to sharp changes in their output.
We describe a set-based framework that successfully deals with real-world CNNs, such as VGG16 and VGG19, that have high accuracy on ImageNet.
arXiv Detail & Related papers (2020-04-12T00:37:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.