Efficient Joint-Dimensional Search with Solution Space Regularization
for Real-Time Semantic Segmentation
- URL: http://arxiv.org/abs/2208.05271v1
- Date: Wed, 10 Aug 2022 11:07:33 GMT
- Title: Efficient Joint-Dimensional Search with Solution Space Regularization
for Real-Time Semantic Segmentation
- Authors: Peng Ye, Baopu Li, Tao Chen, Jiayuan Fan, Zhen Mei, Chen Lin, Chongyan
Zuo, Qinghua Chi, Wanli Ouyan
- Abstract summary: We search an optimal network structure that can run in real-time for this problem.
A novel Solution Space Regularization (SSR) loss is first proposed to effectively encourage the supernet to converge to its discrete one.
A new Hierarchical and Progressive Solution Space Shrinking method is presented to further achieve high efficiency of searching.
- Score: 27.94898516315886
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semantic segmentation is a popular research topic in computer vision, and
many efforts have been made on it with impressive results. In this paper, we
intend to search an optimal network structure that can run in real-time for
this problem. Towards this goal, we jointly search the depth, channel, dilation
rate and feature spatial resolution, which results in a search space consisting
of about 2.78*10^324 possible choices. To handle such a large search space, we
leverage differential architecture search methods. However, the architecture
parameters searched using existing differential methods need to be discretized,
which causes the discretization gap between the architecture parameters found
by the differential methods and their discretized version as the final solution
for the architecture search. Hence, we relieve the problem of discretization
gap from the innovative perspective of solution space regularization.
Specifically, a novel Solution Space Regularization (SSR) loss is first
proposed to effectively encourage the supernet to converge to its discrete one.
Then, a new Hierarchical and Progressive Solution Space Shrinking method is
presented to further achieve high efficiency of searching. In addition, we
theoretically show that the optimization of SSR loss is equivalent to the
L_0-norm regularization, which accounts for the improved search-evaluation gap.
Comprehensive experiments show that the proposed search scheme can efficiently
find an optimal network structure that yields an extremely fast speed (175 FPS)
of segmentation with a small model size (1 M) while maintaining comparable
accuracy.
Related papers
- MGAS: Multi-Granularity Architecture Search for Trade-Off Between Model
Effectiveness and Efficiency [10.641875933652647]
We introduce multi-granularity architecture search (MGAS) to discover both effective and efficient neural networks.
We learn discretization functions specific to each granularity level to adaptively determine the unit remaining ratio according to the evolving architecture.
Extensive experiments on CIFAR-10, CIFAR-100 and ImageNet demonstrate that MGAS outperforms other state-of-the-art methods in achieving a better trade-off between model performance and model size.
arXiv Detail & Related papers (2023-10-23T16:32:18Z) - COIL: Constrained Optimization in Learned Latent Space -- Learning
Representations for Valid Solutions [4.372703857711996]
We use a Variational Autoencoder to learn representations of Constrained Optimization in Latent Space (COIL)
COIL can satisfy constraints and find solutions with distance to objective up to two orders of closer.
We show that, compared to an identical GA using a standard representation, COIL with its learned latent representation can satisfy constraints and find solutions with distance to objective up to two orders of closer.
arXiv Detail & Related papers (2022-02-04T14:45:37Z) - iDARTS: Differentiable Architecture Search with Stochastic Implicit
Gradients [75.41173109807735]
Differentiable ARchiTecture Search (DARTS) has recently become the mainstream of neural architecture search (NAS)
We tackle the hypergradient computation in DARTS based on the implicit function theorem.
We show that the architecture optimisation with the proposed method, named iDARTS, is expected to converge to a stationary point.
arXiv Detail & Related papers (2021-06-21T00:44:11Z) - Towards Improving the Consistency, Efficiency, and Flexibility of
Differentiable Neural Architecture Search [84.4140192638394]
Most differentiable neural architecture search methods construct a super-net for search and derive a target-net as its sub-graph for evaluation.
In this paper, we introduce EnTranNAS that is composed of Engine-cells and Transit-cells.
Our method also spares much memory and computation cost, which speeds up the search process.
arXiv Detail & Related papers (2021-01-27T12:16:47Z) - Trilevel Neural Architecture Search for Efficient Single Image
Super-Resolution [127.92235484598811]
This paper proposes a trilevel neural architecture search (NAS) method for efficient single image super-resolution (SR)
For modeling the discrete search space, we apply a new continuous relaxation on the discrete search spaces to build a hierarchical mixture of network-path, cell-operations, and kernel-width.
An efficient search algorithm is proposed to perform optimization in a hierarchical supernet manner.
arXiv Detail & Related papers (2021-01-17T12:19:49Z) - ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse
Coding [86.40042104698792]
We formulate neural architecture search as a sparse coding problem.
In experiments, our two-stage method on CIFAR-10 requires only 0.05 GPU-day for search.
Our one-stage method produces state-of-the-art performances on both CIFAR-10 and ImageNet at the cost of only evaluation time.
arXiv Detail & Related papers (2020-10-13T04:34:24Z) - Discretization-Aware Architecture Search [81.35557425784026]
This paper presents discretization-aware architecture search (DAtextsuperscript2S)
The core idea is to push the super-network towards the configuration of desired topology, so that the accuracy loss brought by discretization is largely alleviated.
Experiments on standard image classification benchmarks demonstrate the superiority of our approach.
arXiv Detail & Related papers (2020-07-07T01:18:58Z) - DrNAS: Dirichlet Neural Architecture Search [88.56953713817545]
We treat the continuously relaxed architecture mixing weight as random variables, modeled by Dirichlet distribution.
With recently developed pathwise derivatives, the Dirichlet parameters can be easily optimized with gradient-based generalization.
To alleviate the large memory consumption of differentiable NAS, we propose a simple yet effective progressive learning scheme.
arXiv Detail & Related papers (2020-06-18T08:23:02Z) - AlphaGAN: Fully Differentiable Architecture Search for Generative
Adversarial Networks [15.740179244963116]
Generative Adversarial Networks (GANs) are formulated as minimax game problems, whereby generators attempt to approach real data distributions by virtue of adversarial learning against discriminators.
In this work, we aim to boost model learning from the perspective of network architectures, by incorporating recent progress on automated architecture search into GANs.
We propose a fully differentiable search framework for generative adversarial networks, dubbed alphaGAN.
arXiv Detail & Related papers (2020-06-16T13:27:30Z) - MiLeNAS: Efficient Neural Architecture Search via Mixed-Level
Reformulation [25.56562895285528]
mldas is a mixed-level reformulation for NAS that can be optimized efficiently and reliably.
It is shown that even when using a simple first-order method on the mixed-level formulation, mldas can achieve a lower validation error for NAS problems.
arXiv Detail & Related papers (2020-03-27T05:06:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.