When CNNs Meet Random RNNs: Towards Multi-Level Analysis for RGB-D
Object and Scene Recognition
- URL: http://arxiv.org/abs/2004.12349v2
- Date: Tue, 11 Jan 2022 07:41:19 GMT
- Title: When CNNs Meet Random RNNs: Towards Multi-Level Analysis for RGB-D
Object and Scene Recognition
- Authors: Ali Caglayan and Nevrez Imamoglu and Ahmet Burak Can and Ryosuke
Nakamura
- Abstract summary: We propose a novel framework that extracts discriminative feature representations from multi-modal RGB-D images for object and scene recognition tasks.
To cope with the high dimensionality of CNN activations, a random weighted pooling scheme has been proposed.
Experiments verify that fully randomized structure in RNN stage encodes CNN activations to discriminative solid features successfully.
- Score: 10.796613905980609
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recognizing objects and scenes are two challenging but essential tasks in
image understanding. In particular, the use of RGB-D sensors in handling these
tasks has emerged as an important area of focus for better visual
understanding. Meanwhile, deep neural networks, specifically convolutional
neural networks (CNNs), have become widespread and have been applied to many
visual tasks by replacing hand-crafted features with effective deep features.
However, it is an open problem how to exploit deep features from a multi-layer
CNN model effectively. In this paper, we propose a novel two-stage framework
that extracts discriminative feature representations from multi-modal RGB-D
images for object and scene recognition tasks. In the first stage, a pretrained
CNN model has been employed as a backbone to extract visual features at
multiple levels. The second stage maps these features into high level
representations with a fully randomized structure of recursive neural networks
(RNNs) efficiently. To cope with the high dimensionality of CNN activations, a
random weighted pooling scheme has been proposed by extending the idea of
randomness in RNNs. Multi-modal fusion has been performed through a soft voting
approach by computing weights based on individual recognition confidences (i.e.
SVM scores) of RGB and depth streams separately. This produces consistent class
label estimation in final RGB-D classification performance. Extensive
experiments verify that fully randomized structure in RNN stage encodes CNN
activations to discriminative solid features successfully. Comparative
experimental results on the popular Washington RGB-D Object and SUN RGB-D Scene
datasets show that the proposed approach achieves superior or on-par
performance compared to state-of-the-art methods both in object and scene
recognition tasks. Code is available at
https://github.com/acaglayan/CNN_randRNN.
Related papers
- Kronecker Product Feature Fusion for Convolutional Neural Network in
Remote Sensing Scene Classification [0.0]
CNN can extract hierarchical convolutional features from remote sensing imagery.
Two successful Feature Fusion methods, Add and Concat, are employed in certain state-of-the-art CNN algorithms.
We propose a novel Feature Fusion algorithm, which unifies the aforementioned methods using the Kronecker Product (KPFF)
arXiv Detail & Related papers (2024-01-08T19:01:01Z) - Point-aware Interaction and CNN-induced Refinement Network for RGB-D
Salient Object Detection [95.84616822805664]
We introduce CNNs-assisted Transformer architecture and propose a novel RGB-D SOD network with Point-aware Interaction and CNN-induced Refinement.
In order to alleviate the block effect and detail destruction problems brought by the Transformer naturally, we design a CNN-induced refinement (CNNR) unit for content refinement and supplementation.
arXiv Detail & Related papers (2023-08-17T11:57:49Z) - Neural Implicit Dictionary via Mixture-of-Expert Training [111.08941206369508]
We present a generic INR framework that achieves both data and training efficiency by learning a Neural Implicit Dictionary (NID)
Our NID assembles a group of coordinate-based Impworks which are tuned to span the desired function space.
Our experiments show that, NID can improve reconstruction of 2D images or 3D scenes by 2 orders of magnitude faster with up to 98% less input data.
arXiv Detail & Related papers (2022-07-08T05:07:19Z) - Depth-Adapted CNNs for RGB-D Semantic Segmentation [2.341385717236931]
We propose a novel framework to incorporate the depth information in the RGB convolutional neural network (CNN)
Specifically, our Z-ACN generates a 2D depth-adapted offset which is fully constrained by low-level features to guide the feature extraction on RGB images.
With the generated offset, we introduce two intuitive and effective operations to replace basic CNN operators.
arXiv Detail & Related papers (2022-06-08T14:59:40Z) - A Novel Hand Gesture Detection and Recognition system based on
ensemble-based Convolutional Neural Network [3.5665681694253903]
Detection of hand portion has become a challenging task in computer vision and pattern recognition communities.
Deep learning algorithm like convolutional neural network (CNN) architecture has become a very popular choice for classification tasks.
In this paper, an ensemble of CNN-based approaches is presented to overcome some problems like high variance during prediction, overfitting problem and also prediction errors.
arXiv Detail & Related papers (2022-02-25T06:46:58Z) - RGB-D SLAM Using Attention Guided Frame Association [11.484398586420067]
We propose the use of task specific network attention for RGB-D indoor SLAM.
We integrate layer-wise object attention information (layer gradients) with CNN layer representations to improve frame association performance.
Experiments show promising initial results with improved performance.
arXiv Detail & Related papers (2022-01-28T11:23:29Z) - Hybrid SNN-ANN: Energy-Efficient Classification and Object Detection for
Event-Based Vision [64.71260357476602]
Event-based vision sensors encode local pixel-wise brightness changes in streams of events rather than image frames.
Recent progress in object recognition from event-based sensors has come from conversions of deep neural networks.
We propose a hybrid architecture for end-to-end training of deep neural networks for event-based pattern recognition and object detection.
arXiv Detail & Related papers (2021-12-06T23:45:58Z) - New SAR target recognition based on YOLO and very deep multi-canonical
correlation analysis [0.1503974529275767]
This paper proposes a robust feature extraction method for SAR image target classification by adaptively fusing effective features from different CNN layers.
Experiments on the MSTAR dataset demonstrate that the proposed method outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2021-10-28T18:10:26Z) - HAT: Hierarchical Aggregation Transformers for Person Re-identification [87.02828084991062]
We take advantages of both CNNs and Transformers for image-based person Re-ID with high performance.
Work is the first to take advantages of both CNNs and Transformers for image-based person Re-ID.
arXiv Detail & Related papers (2021-07-13T09:34:54Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection [76.30585706811993]
We present a novel and high-performance 3D object detection framework, named PointVoxel-RCNN (PV-RCNN)
Our proposed method deeply integrates both 3D voxel Convolutional Neural Network (CNN) and PointNet-based set abstraction.
It takes advantages of efficient learning and high-quality proposals of the 3D voxel CNN and the flexible receptive fields of the PointNet-based networks.
arXiv Detail & Related papers (2019-12-31T06:34:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.