Toward Edge-Efficient Dense Predictions with Synergistic Multi-Task
Neural Architecture Search
- URL: http://arxiv.org/abs/2210.01384v1
- Date: Tue, 4 Oct 2022 04:49:08 GMT
- Title: Toward Edge-Efficient Dense Predictions with Synergistic Multi-Task
Neural Architecture Search
- Authors: Thanh Vu, Yanqi Zhou, Chunfeng Wen, Yueqi Li, Jan-Michael Frahm
- Abstract summary: We propose a novel and scalable solution to address the challenges of developing efficient dense predictions on edge platforms.
Our first key insight is that MultiTask Learning (MTL) and hardware-aware Neural Architecture Search (NAS) can work in synergy to greatly benefit on-device Dense Predictions (DP)
We propose JAReD, an improved, easy-to-adopt Joint Absolute-Relative Depth loss, that reduces up to 88% of the undesired noise while simultaneously boosting accuracy.
- Score: 22.62389136288258
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we propose a novel and scalable solution to address the
challenges of developing efficient dense predictions on edge platforms. Our
first key insight is that MultiTask Learning (MTL) and hardware-aware Neural
Architecture Search (NAS) can work in synergy to greatly benefit on-device
Dense Predictions (DP). Empirical results reveal that the joint learning of the
two paradigms is surprisingly effective at improving DP accuracy, achieving
superior performance over both the transfer learning of single-task NAS and
prior state-of-the-art approaches in MTL, all with just 1/10th of the
computation. To the best of our knowledge, our framework, named EDNAS, is the
first to successfully leverage the synergistic relationship of NAS and MTL for
DP. Our second key insight is that the standard depth training for multi-task
DP can cause significant instability and noise to MTL evaluation. Instead, we
propose JAReD, an improved, easy-to-adopt Joint Absolute-Relative Depth loss,
that reduces up to 88% of the undesired noise while simultaneously boosting
accuracy. We conduct extensive evaluations on standard datasets, benchmark
against strong baselines and state-of-the-art approaches, as well as provide an
analysis of the discovered optimal architectures.
Related papers
- POMONAG: Pareto-Optimal Many-Objective Neural Architecture Generator [4.09225917049674]
Transferable NAS has emerged, generalizing the search process from dataset-dependent to task-dependent.
This paper introduces POMONAG, extending DiffusionNAG via a many-optimal diffusion process.
Results were validated on two search spaces -- NAS201 and MobileNetV3 -- and evaluated across 15 image classification datasets.
arXiv Detail & Related papers (2024-09-30T16:05:29Z) - A Pairwise Comparison Relation-assisted Multi-objective Evolutionary Neural Architecture Search Method with Multi-population Mechanism [58.855741970337675]
Neural architecture search (NAS) enables re-searchers to automatically explore vast search spaces and find efficient neural networks.
NAS suffers from a key bottleneck, i.e., numerous architectures need to be evaluated during the search process.
We propose the SMEM-NAS, a pairwise com-parison relation-assisted multi-objective evolutionary algorithm based on a multi-population mechanism.
arXiv Detail & Related papers (2024-07-22T12:46:22Z) - Intuition-aware Mixture-of-Rank-1-Experts for Parameter Efficient Finetuning [50.73666458313015]
Large Language Models (LLMs) have demonstrated significant potential in performing multiple tasks in multimedia applications.
MoE has been emerged as a promising solution with its sparse architecture for effective task decoupling.
Intuition-MoR1E achieves superior efficiency and 2.15% overall accuracy improvement across 14 public datasets.
arXiv Detail & Related papers (2024-04-13T12:14:58Z) - Fast Neural Architecture Search for Lightweight Dense Prediction
Networks [41.605107921584775]
We present LDP, a lightweight dense prediction neural architecture search (NAS) framework.
Starting from a pre-defined generic backbone, LDP applies the novel Assisted Tabu Search for efficient architecture exploration.
Experiments show that the proposed framework yields consistent improvements on all tested dense prediction tasks.
arXiv Detail & Related papers (2022-03-03T20:17:10Z) - ZARTS: On Zero-order Optimization for Neural Architecture Search [94.41017048659664]
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency.
This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation.
In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue.
arXiv Detail & Related papers (2021-10-10T09:35:15Z) - CM-NAS: Cross-Modality Neural Architecture Search for Visible-Infrared
Person Re-Identification [102.89434996930387]
VI-ReID aims to match cross-modality pedestrian images, breaking through the limitation of single-modality person ReID in dark environment.
Existing works manually design various two-stream architectures to separately learn modality-specific and modality-sharable representations.
We propose a novel method, named Cross-Modality Neural Architecture Search (CM-NAS)
arXiv Detail & Related papers (2021-01-21T07:07:00Z) - EQ-Net: A Unified Deep Learning Framework for Log-Likelihood Ratio
Estimation and Quantization [25.484585922608193]
We introduce EQ-Net: the first holistic framework that solves both the tasks of log-likelihood ratio (LLR) estimation and quantization using a data-driven method.
We carry out extensive experimental evaluation and demonstrate that our single architecture achieves state-of-the-art results on both tasks.
arXiv Detail & Related papers (2020-12-23T18:11:30Z) - Towards Accurate Knowledge Transfer via Target-awareness Representation
Disentanglement [56.40587594647692]
We propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED)
TRED disentangles the relevant knowledge with respect to the target task from the original source model and used as a regularizer during fine-tuning the target model.
Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2% in average.
arXiv Detail & Related papers (2020-10-16T17:45:08Z) - Generalized Reinforcement Meta Learning for Few-Shot Optimization [3.7675996866306845]
We present a generic and flexible Reinforcement Learning (RL) based meta-learning framework for the problem of few-shot learning.
Our framework could be easily extended to do network architecture search.
arXiv Detail & Related papers (2020-05-04T03:21:05Z) - Multi-Task Learning for Dense Prediction Tasks: A Survey [87.66280582034838]
Multi-task learning (MTL) techniques have shown promising results w.r.t. performance, computations and/or memory footprint.
We provide a well-rounded view on state-of-the-art deep learning approaches for MTL in computer vision.
arXiv Detail & Related papers (2020-04-28T09:15:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.