Representation Sharing for Fast Object Detector Search and Beyond
- URL: http://arxiv.org/abs/2007.12075v4
- Date: Fri, 23 Oct 2020 07:55:42 GMT
- Title: Representation Sharing for Fast Object Detector Search and Beyond
- Authors: Yujie Zhong, Zelu Deng, Sheng Guo, Matthew R. Scott, Weilin Huang
- Abstract summary: We propose Fast And Diverse (FAD) to better explore the optimal configuration of receptive fields and convolution types in the sub-networks for one-stage detectors.
FAD achieves prominent improvements on two types of one-stage detectors with various backbones.
- Score: 38.18583590914755
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Region Proposal Network (RPN) provides strong support for handling the scale
variation of objects in two-stage object detection. For one-stage detectors
which do not have RPN, it is more demanding to have powerful sub-networks
capable of directly capturing objects of unknown sizes. To enhance such
capability, we propose an extremely efficient neural architecture search
method, named Fast And Diverse (FAD), to better explore the optimal
configuration of receptive fields and convolution types in the sub-networks for
one-stage detectors. FAD consists of a designed search space and an efficient
architecture search algorithm. The search space contains a rich set of diverse
transformations designed specifically for object detection. To cope with the
designed search space, a novel search algorithm termed Representation Sharing
(RepShare) is proposed to effectively identify the best combinations of the
defined transformations. In our experiments, FAD obtains prominent improvements
on two types of one-stage detectors with various backbones. In particular, our
FAD detector achieves 46.4 AP on MS-COCO (under single-scale testing),
outperforming the state-of-the-art detectors, including the most recent
NAS-based detectors, Auto-FPN (searched for 16 GPU-days) and NAS-FCOS (28
GPU-days), while significantly reduces the search cost to 0.6 GPU-days. Beyond
object detection, we further demonstrate the generality of FAD on the more
challenging instance segmentation, and expect it to benefit more tasks.
Related papers
- Renormalized Connection for Scale-preferred Object Detection in Satellite Imagery [51.83786195178233]
We design a Knowledge Discovery Network (KDN) to implement the renormalization group theory in terms of efficient feature extraction.
Renormalized connection (RC) on the KDN enables synergistic focusing'' of multi-scale features.
RCs extend the multi-level feature's divide-and-conquer'' mechanism of the FPN-based detectors to a wide range of scale-preferred tasks.
arXiv Detail & Related papers (2024-09-09T13:56:22Z) - Multi-Objective Evolutionary for Object Detection Mobile Architectures
Search [21.14296703753317]
We propose a mobile object detection backbone network architecture search algorithm based on non-dominated sorting for NAS scenarios.
The proposed approach can search the backbone networks with different depths, widths, or expansion sizes via a technique of weight mapping.
Under similar computational complexity, the accuracy of the backbone network architecture we search for is 2.0% mAP higher than MobileDet.
arXiv Detail & Related papers (2022-11-05T00:28:49Z) - Searching a High-Performance Feature Extractor for Text Recognition
Network [92.12492627169108]
We design a domain-specific search space by exploring principles for having good feature extractors.
As the space is huge and complexly structured, no existing NAS algorithms can be applied.
We propose a two-stage algorithm to effectively search in the space.
arXiv Detail & Related papers (2022-09-27T03:49:04Z) - EAutoDet: Efficient Architecture Search for Object Detection [110.99532343155073]
EAutoDet framework can discover practical backbone and FPN architectures for object detection in 1.4 GPU-days.
We propose a kernel reusing technique by sharing the weights of candidate operations on one edge and consolidating them into one convolution.
In particular, the discovered architectures surpass state-of-the-art object detection NAS methods and achieve 40.1 mAP with 120 FPS and 49.2 mAP with 41.3 FPS on COCO test-dev set.
arXiv Detail & Related papers (2022-03-21T05:56:12Z) - MRDet: A Multi-Head Network for Accurate Oriented Object Detection in
Aerial Images [51.227489316673484]
We propose an arbitrary-oriented region proposal network (AO-RPN) to generate oriented proposals transformed from horizontal anchors.
To obtain accurate bounding boxes, we decouple the detection task into multiple subtasks and propose a multi-head network.
Each head is specially designed to learn the features optimal for the corresponding task, which allows our network to detect objects accurately.
arXiv Detail & Related papers (2020-12-24T06:36:48Z) - AFD-Net: Adaptive Fully-Dual Network for Few-Shot Object Detection [8.39479809973967]
Few-shot object detection (FSOD) aims at learning a detector that can fast adapt to previously unseen objects with scarce examples.
Existing methods solve this problem by performing subtasks of classification and localization utilizing a shared component.
We present that a general few-shot detector should consider the explicit decomposition of two subtasks, as well as leveraging information from both of them to enhance feature representations.
arXiv Detail & Related papers (2020-11-30T10:21:32Z) - Hit-Detector: Hierarchical Trinity Architecture Search for Object
Detection [67.84976857449263]
We propose a hierarchical trinity search framework to simultaneously discover efficient architectures for all components of object detector.
We employ a novel scheme to automatically screen different sub search spaces for different components so as to perform the end-to-end search for each component efficiently.
Our searched architecture, namely Hit-Detector, achieves 41.4% mAP on COCO minival set with 27M parameters.
arXiv Detail & Related papers (2020-03-26T10:20:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.