Learning to Generate Content-Aware Dynamic Detectors
- URL: http://arxiv.org/abs/2012.04265v1
- Date: Tue, 8 Dec 2020 08:05:20 GMT
- Title: Learning to Generate Content-Aware Dynamic Detectors
- Authors: Junyi Feng, Jiashen Hua, Baisheng Lai, Jianqiang Huang, Xi Li,
Xian-sheng Hua
- Abstract summary: We introduce a newpective of designing efficient detectors, which is automatically generating sample-adaptive model architecture.
We introduce a course-to-fine strat-egy tailored for object detection to guide the learning of dynamic routing.
Experiments on MS-COCO dataset demonstrate that CADDet achieves 1.8 higher mAP with 10% fewer FLOPs compared with vanilla routing.
- Score: 62.74209921174237
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model efficiency is crucial for object detection. Mostprevious works rely on
either hand-crafted design or auto-search methods to obtain a static
architecture, regardless ofthe difference of inputs. In this paper, we
introduce a newperspective of designing efficient detectors, which is
automatically generating sample-adaptive model architectureon the fly. The
proposed method is named content-aware dynamic detectors (CADDet). It first
applies a multi-scale densely connected network with dynamic routing as the
supernet. Furthermore, we introduce a course-to-fine strat-egy tailored for
object detection to guide the learning of dynamic routing, which contains two
metrics: 1) dynamic global budget constraint assigns data-dependent
expectedbudgets for individual samples; 2) local path similarity regularization
aims to generate more diverse routing paths. With these, our method achieves
higher computational efficiency while maintaining good performance. To the best
of our knowledge, our CADDet is the first work to introduce dynamic routing
mechanism in object detection. Experiments on MS-COCO dataset demonstrate that
CADDet achieves 1.8 higher mAP with 10% fewer FLOPs compared with vanilla
routing strategy. Compared with the models based upon similar building blocks,
CADDet achieves a 42% FLOPs reduction with a competitive mAP.
Related papers
- Dynamic Perceiver for Efficient Visual Recognition [87.08210214417309]
We propose Dynamic Perceiver (Dyn-Perceiver) to decouple the feature extraction procedure and the early classification task.
A feature branch serves to extract image features, while a classification branch processes a latent code assigned for classification tasks.
Early exits are placed exclusively within the classification branch, thus eliminating the need for linear separability in low-level features.
arXiv Detail & Related papers (2023-06-20T03:00:22Z) - DynamicDet: A Unified Dynamic Architecture for Object Detection [9.719671347009827]
We propose a dynamic framework for object detection, named DynamicDet.
We present a novel optimization strategy with an exiting criterion based on the detection losses for our dynamic detectors.
Experiments conducted on the COCO benchmark demonstrate that the proposed DynamicDet achieves new state-of-the-art accuracy-speed trade-offs.
arXiv Detail & Related papers (2023-04-12T01:16:53Z) - Boosting the Efficiency of Parametric Detection with Hierarchical Neural
Networks [4.1410005218338695]
We propose Hierarchical Detection Network (HDN), a novel approach to efficient detection.
The network is trained using a novel loss function, which encodes simultaneously the goals of statistical accuracy and efficiency.
We show how training a three-layer HDN using two-layer model can further boost both accuracy and efficiency.
arXiv Detail & Related papers (2022-07-23T19:23:00Z) - Joint Spatial-Temporal and Appearance Modeling with Transformer for
Multiple Object Tracking [59.79252390626194]
We propose a novel solution named TransSTAM, which leverages Transformer to model both the appearance features of each object and the spatial-temporal relationships among objects.
The proposed method is evaluated on multiple public benchmarks including MOT16, MOT17, and MOT20, and it achieves a clear performance improvement in both IDF1 and HOTA.
arXiv Detail & Related papers (2022-05-31T01:19:18Z) - DAAS: Differentiable Architecture and Augmentation Policy Search [107.53318939844422]
This work considers the possible coupling between neural architectures and data augmentation and proposes an effective algorithm jointly searching for them.
Our approach achieves 97.91% accuracy on CIFAR-10 and 76.6% Top-1 accuracy on ImageNet dataset, showing the outstanding performance of our search algorithm.
arXiv Detail & Related papers (2021-09-30T17:15:17Z) - DSIC: Dynamic Sample-Individualized Connector for Multi-Scale Object
Detection [33.61001547745264]
We propose a Dynamic Sample-Individualized Connector (DSIC) for multi-scale object detection.
ISG adaptively extracts multi-level features from backbone as the input of feature integration.
CSG automatically activate informative data flow paths based on the multi-level features.
arXiv Detail & Related papers (2020-11-16T08:13:58Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z) - Learning Dynamic Routing for Semantic Segmentation [86.56049245100084]
This paper studies a conceptually new method to alleviate the scale variance in semantic representation, named dynamic routing.
The proposed framework generates data-dependent routes, adapting to the scale distribution of each image.
To this end, a differentiable gating function, called soft conditional gate, is proposed to select scale transform paths on the fly.
arXiv Detail & Related papers (2020-03-23T17:22:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.