Designing Deep Networks for Scene Recognition
- URL: http://arxiv.org/abs/2303.07402v1
- Date: Mon, 13 Mar 2023 18:28:06 GMT
- Title: Designing Deep Networks for Scene Recognition
- Authors: Zhinan Qiao, Xiaohui Yuan
- Abstract summary: We conduct extensive experiments to demonstrate the widely accepted principles in network design may result in dramatic performance differences when the data is altered.
This paper presents a novel network design methodology: data-oriented network design.
We propose a Deep-Narrow Network and Dilated Pooling module, which improved the scene recognition performance using less than half of the computational resources.
- Score: 3.493180651702109
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most deep learning backbones are evaluated on ImageNet. Using scenery images
as an example, we conducted extensive experiments to demonstrate the widely
accepted principles in network design may result in dramatic performance
differences when the data is altered. Exploratory experiments are engaged to
explain the underlining cause of the differences. Based on our observation,
this paper presents a novel network design methodology: data-oriented network
design. In other words, instead of designing universal backbones, the scheming
of the networks should treat the characteristics of data as a crucial
component. We further proposed a Deep-Narrow Network and Dilated Pooling
module, which improved the scene recognition performance using less than half
of the computational resources compared to the benchmark network architecture
ResNets. The source code is publicly available on
https://github.com/ZN-Qiao/Deep-Narrow-Network.
Related papers
- Efficient and Accurate Hyperspectral Image Demosaicing with Neural Network Architectures [3.386560551295746]
This study investigates the effectiveness of neural network architectures in hyperspectral image demosaicing.
We introduce a range of network models and modifications, and compare them with classical methods and existing reference network approaches.
Results indicate that our networks outperform or match reference models in both datasets demonstrating exceptional performance.
arXiv Detail & Related papers (2023-12-21T08:02:49Z) - Visual Prompting Upgrades Neural Network Sparsification: A Data-Model Perspective [64.04617968947697]
We introduce a novel data-model co-design perspective: to promote superior weight sparsity.
Specifically, customized Visual Prompts are mounted to upgrade neural Network sparsification in our proposed VPNs framework.
arXiv Detail & Related papers (2023-12-03T13:50:24Z) - Impact of Scaled Image on Robustness of Deep Neural Networks [0.0]
Scaling the raw images creates out-of-distribution data, which makes it a possible adversarial attack to fool the networks.
In this work, we propose a Scaling-distortion dataset ImageNet-CS by Scaling a subset of the ImageNet Challenge dataset by different multiples.
arXiv Detail & Related papers (2022-09-02T08:06:58Z) - Learning Deep Context-Sensitive Decomposition for Low-Light Image
Enhancement [58.72667941107544]
A typical framework is to simultaneously estimate the illumination and reflectance, but they disregard the scene-level contextual information encapsulated in feature spaces.
We develop a new context-sensitive decomposition network architecture to exploit the scene-level contextual dependencies on spatial scales.
We develop a lightweight CSDNet (named LiteCSDNet) by reducing the number of channels.
arXiv Detail & Related papers (2021-12-09T06:25:30Z) - Smart Data Representations: Impact on the Accuracy of Deep Neural
Networks [0.2446672595462589]
We analyze the impact of data representations on the performance of Deep Neural Networks using energy time series forecasting.
The results show that, depending on the forecast horizon, the same data representations can have a positive or negative impact on the accuracy of Deep Neural Networks.
arXiv Detail & Related papers (2021-11-17T14:06:08Z) - Recurrence along Depth: Deep Convolutional Neural Networks with
Recurrent Layer Aggregation [5.71305698739856]
This paper introduces a concept of layer aggregation to describe how information from previous layers can be reused to better extract features at the current layer.
We propose a very light-weighted module, called recurrent layer aggregation (RLA), by making use of the sequential structure of layers in a deep CNN.
Our RLA module is compatible with many mainstream deep CNNs, including ResNets, Xception and MobileNetV2.
arXiv Detail & Related papers (2021-10-22T15:36:33Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z) - The Heterogeneity Hypothesis: Finding Layer-Wise Differentiated Network
Architectures [179.66117325866585]
We investigate a design space that is usually overlooked, i.e. adjusting the channel configurations of predefined networks.
We find that this adjustment can be achieved by shrinking widened baseline networks and leads to superior performance.
Experiments are conducted on various networks and datasets for image classification, visual tracking and image restoration.
arXiv Detail & Related papers (2020-06-29T17:59:26Z) - Boosting Connectivity in Retinal Vessel Segmentation via a Recursive
Semantics-Guided Network [23.936946593048987]
A U-shape network is enhanced by introducing a semantics-guided module, which integrates the enriched semantics information to shallow layers for guiding the network to explore more powerful features.
The carefully designed semantics-guided network has been extensively evaluated on several public datasets.
arXiv Detail & Related papers (2020-04-24T09:18:04Z) - ResNeSt: Split-Attention Networks [86.25490825631763]
We present a modularized architecture, which applies the channel-wise attention on different network branches to leverage their success in capturing cross-feature interactions and learning diverse representations.
Our model, named ResNeSt, outperforms EfficientNet in accuracy and latency trade-off on image classification.
arXiv Detail & Related papers (2020-04-19T20:40:31Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.