RSBNet: One-Shot Neural Architecture Search for A Backbone Network in
Remote Sensing Image Recognition
- URL: http://arxiv.org/abs/2112.03456v1
- Date: Tue, 7 Dec 2021 02:44:16 GMT
- Title: RSBNet: One-Shot Neural Architecture Search for A Backbone Network in
Remote Sensing Image Recognition
- Authors: Cheng Peng, Yangyang Li, Ronghua Shang, Licheng Jiao
- Abstract summary: We propose a new design paradigm for the backbone architecture in RSI recognition tasks, including scene classification, land-cover classification and object detection.
A novel one-shot architecture search framework based on weight-sharing strategy and evolutionary algorithm is proposed, called RSBNet.
Extensive experiments have been conducted on five benchmark datasets for different recognition tasks, the results show the effectiveness of the proposed search paradigm.
- Score: 43.95699860302204
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, a massive number of deep learning based approaches have been
successfully applied to various remote sensing image (RSI) recognition tasks.
However, most existing advances of deep learning methods in the RSI field
heavily rely on the features extracted by the manually designed backbone
network, which severely hinders the potential of deep learning models due the
complexity of RSI and the limitation of prior knowledge. In this paper, we
research a new design paradigm for the backbone architecture in RSI recognition
tasks, including scene classification, land-cover classification and object
detection. A novel one-shot architecture search framework based on
weight-sharing strategy and evolutionary algorithm is proposed, called RSBNet,
which consists of three stages: Firstly, a supernet constructed in a layer-wise
search space is pretrained on a self-assembled large-scale RSI dataset based on
an ensemble single-path training strategy. Next, the pre-trained supernet is
equipped with different recognition heads through the switchable recognition
module and respectively fine-tuned on the target dataset to obtain
task-specific supernet. Finally, we search the optimal backbone architecture
for different recognition tasks based on the evolutionary algorithm without any
network training. Extensive experiments have been conducted on five benchmark
datasets for different recognition tasks, the results show the effectiveness of
the proposed search paradigm and demonstrate that the searched backbone is able
to flexibly adapt different RSI recognition tasks and achieve impressive
performance.
Related papers
- EM-DARTS: Hierarchical Differentiable Architecture Search for Eye Movement Recognition [54.99121380536659]
Eye movement biometrics have received increasing attention thanks to its high secure identification.
Deep learning (DL) models have been recently successfully applied for eye movement recognition.
DL architecture still is determined by human prior knowledge.
We propose EM-DARTS, a hierarchical differentiable architecture search algorithm to automatically design the DL architecture for eye movement recognition.
arXiv Detail & Related papers (2024-09-22T13:11:08Z) - Designing deep neural networks for driver intention recognition [40.87622566719826]
This paper applies neural architecture search to investigate the effects of the deep neural network architecture on a real-world safety critical application.
A set of eight search strategies are evaluated for two driver intention recognition datasets.
arXiv Detail & Related papers (2024-02-07T12:54:15Z) - SIRe-Networks: Skip Connections over Interlaced Multi-Task Learning and
Residual Connections for Structure Preserving Object Classification [28.02302915971059]
In this paper, we introduce an interlaced multi-task learning strategy, defined SIRe, to reduce the vanishing gradient in relation to the object classification task.
The presented methodology directly improves a convolutional neural network (CNN) by enforcing the input image structure preservation through auto-encoders.
To validate the presented methodology, a simple CNN and various implementations of famous networks are extended via the SIRe strategy and extensively tested on the CIFAR100 dataset.
arXiv Detail & Related papers (2021-10-06T13:54:49Z) - Differentiable Neural Architecture Search with Morphism-based
Transformable Backbone Architectures [35.652234989200956]
This study aims at making the architecture search process more adaptive for one-shot or online training.
It introduces a growing mechanism for differentiable neural architecture search based on network morphism.
We also implement a recently proposed two-input backbone architecture for recurrent neural networks.
arXiv Detail & Related papers (2021-06-14T07:56:33Z) - Joint Learning of Neural Transfer and Architecture Adaptation for Image
Recognition [77.95361323613147]
Current state-of-the-art visual recognition systems rely on pretraining a neural network on a large-scale dataset and finetuning the network weights on a smaller dataset.
In this work, we prove that dynamically adapting network architectures tailored for each domain task along with weight finetuning benefits in both efficiency and effectiveness.
Our method can be easily generalized to an unsupervised paradigm by replacing supernet training with self-supervised learning in the source domain tasks and performing linear evaluation in the downstream tasks.
arXiv Detail & Related papers (2021-03-31T08:15:17Z) - SAR-NAS: Skeleton-based Action Recognition via Neural Architecture
Searching [18.860051578038608]
We encode a skeleton-based action instance into a tensor and define a set of operations to build two types of network cells: normal cells and reduction cells.
Experiments on the challenging NTU RGB+D and Kinectics datasets have verified that most of the networks developed to date for skeleton-based action recognition are likely not compact and efficient.
The proposed method provides an approach to search for such a compact network that is able to achieve comparative or even better performance than the state-of-the-art methods.
arXiv Detail & Related papers (2020-10-29T03:24:15Z) - NAS-Navigator: Visual Steering for Explainable One-Shot Deep Neural
Network Synthesis [53.106414896248246]
We present a framework that allows analysts to effectively build the solution sub-graph space and guide the network search by injecting their domain knowledge.
Applying this technique in an iterative manner allows analysts to converge to the best performing neural network architecture for a given application.
arXiv Detail & Related papers (2020-09-28T01:48:45Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z) - Deep Multimodal Neural Architecture Search [178.35131768344246]
We devise a generalized deep multimodal neural architecture search (MMnas) framework for various multimodal learning tasks.
Given multimodal input, we first define a set of primitive operations, and then construct a deep encoder-decoder based unified backbone.
On top of the unified backbone, we attach task-specific heads to tackle different multimodal learning tasks.
arXiv Detail & Related papers (2020-04-25T07:00:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.