Joint-DetNAS: Upgrade Your Detector with NAS, Pruning and Dynamic
Distillation
- URL: http://arxiv.org/abs/2105.12971v1
- Date: Thu, 27 May 2021 07:25:43 GMT
- Title: Joint-DetNAS: Upgrade Your Detector with NAS, Pruning and Dynamic
Distillation
- Authors: Lewei Yao, Renjie Pi, Hang Xu, Wei Zhang, Zhenguo Li, Tong Zhang
- Abstract summary: We propose Joint-DetNAS, a unified NAS framework for object detection.
Joint-DetNAS integrates 3 key components: Neural Architecture Search, pruning, and Knowledge Distillation.
Our algorithm directly outputs the derived student detector with high performance without additional training.
- Score: 49.421099172544196
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose Joint-DetNAS, a unified NAS framework for object detection, which
integrates 3 key components: Neural Architecture Search, pruning, and Knowledge
Distillation. Instead of naively pipelining these techniques, our Joint-DetNAS
optimizes them jointly. The algorithm consists of two core processes: student
morphism optimizes the student's architecture and removes the redundant
parameters, while dynamic distillation aims to find the optimal matching
teacher. For student morphism, weight inheritance strategy is adopted, allowing
the student to flexibly update its architecture while fully utilize the
predecessor's weights, which considerably accelerates the search; To facilitate
dynamic distillation, an elastic teacher pool is trained via integrated
progressive shrinking strategy, from which teacher detectors can be sampled
without additional cost in subsequent searches. Given a base detector as the
input, our algorithm directly outputs the derived student detector with high
performance without additional training. Experiments demonstrate that our
Joint-DetNAS outperforms the naive pipelining approach by a great margin. Given
a classic R101-FPN as the base detector, Joint-DetNAS is able to boost its mAP
from 41.4 to 43.9 on MS COCO and reduce the latency by 47%, which is on par
with the SOTA EfficientDet while requiring less search cost. We hope our
proposed method can provide the community with a new way of jointly optimizing
NAS, KD and pruning.
Related papers
- Lightweight Diffusion Models with Distillation-Based Block Neural
Architecture Search [55.41583104734349]
We propose to automatically remove structural redundancy in diffusion models with our proposed Diffusion Distillation-based Block-wise Neural Architecture Search (NAS)
Given a larger pretrained teacher, we leverage DiffNAS to search for the smallest architecture which can achieve on-par or even better performance than the teacher.
Different from previous block-wise NAS methods, DiffNAS contains a block-wise local search strategy and a retraining strategy with a joint dynamic loss.
arXiv Detail & Related papers (2023-11-08T12:56:59Z) - DCP-NAS: Discrepant Child-Parent Neural Architecture Search for 1-bit
CNNs [53.82853297675979]
1-bit convolutional neural networks (CNNs) with binary weights and activations show their potential for resource-limited embedded devices.
One natural approach is to use 1-bit CNNs to reduce the computation and memory cost of NAS.
We introduce Discrepant Child-Parent Neural Architecture Search (DCP-NAS) to efficiently search 1-bit CNNs.
arXiv Detail & Related papers (2023-06-27T11:28:29Z) - $\beta$-DARTS: Beta-Decay Regularization for Differentiable Architecture
Search [85.84110365657455]
We propose a simple-but-efficient regularization method, termed as Beta-Decay, to regularize the DARTS-based NAS searching process.
Experimental results on NAS-Bench-201 show that our proposed method can help to stabilize the searching process and makes the searched network more transferable across different datasets.
arXiv Detail & Related papers (2022-03-03T11:47:14Z) - L$^{2}$NAS: Learning to Optimize Neural Architectures via
Continuous-Action Reinforcement Learning [23.25155249879658]
Differentiable architecture search (NAS) achieved remarkable results in deep neural network design.
We show that L$2$ achieves state-of-theart results on DART201 benchmark as well as NASS and Once-for-All search policies.
arXiv Detail & Related papers (2021-09-25T19:26:30Z) - OPANAS: One-Shot Path Aggregation Network Architecture Search for Object
Detection [82.04372532783931]
Recently, neural architecture search (NAS) has been exploited to design feature pyramid networks (FPNs)
We propose a novel One-Shot Path Aggregation Network Architecture Search (OPANAS) algorithm, which significantly improves both searching efficiency and detection accuracy.
arXiv Detail & Related papers (2021-03-08T01:48:53Z) - DrNAS: Dirichlet Neural Architecture Search [88.56953713817545]
We treat the continuously relaxed architecture mixing weight as random variables, modeled by Dirichlet distribution.
With recently developed pathwise derivatives, the Dirichlet parameters can be easily optimized with gradient-based generalization.
To alleviate the large memory consumption of differentiable NAS, we propose a simple yet effective progressive learning scheme.
arXiv Detail & Related papers (2020-06-18T08:23:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.