Coarse-to-Fine Searching for Efficient Generative Adversarial Networks
- URL: http://arxiv.org/abs/2104.09223v1
- Date: Mon, 19 Apr 2021 11:46:20 GMT
- Title: Coarse-to-Fine Searching for Efficient Generative Adversarial Networks
- Authors: Jiahao Wang, Han Shu, Weihao Xia, Yujiu Yang, Yunhe Wang
- Abstract summary: generative adversarial network (GAN) are usually designed to conduct various complex image generation.
We first discover an intact search space of generator networks including three dimensionalities, i.e., path, operator, channel for fully excavating the network performance.
To reduce the huge search cost, we explore a coarse-to-fine search strategy which divides the overall search process into three sub-optimization problems accordingly.
- Score: 43.21560798088658
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies the neural architecture search (NAS) problem for
developing efficient generator networks. Compared with deep models for visual
recognition tasks, generative adversarial network (GAN) are usually designed to
conduct various complex image generation. We first discover an intact search
space of generator networks including three dimensionalities, i.e., path,
operator, channel for fully excavating the network performance. To reduce the
huge search cost, we explore a coarse-to-fine search strategy which divides the
overall search process into three sub-optimization problems accordingly. In
addition, a fair supernet training approach is utilized to ensure that all
sub-networks can be updated fairly and stably. Experiments results on
benchmarks show that we can provide generator networks with better image
quality and lower computational costs over the state-of-the-art methods. For
example, with our method, it takes only about 8 GPU hours on the entire
edges-to-shoes dataset to get a 2.56 MB model with a 24.13 FID score and 10 GPU
hours on the entire Urban100 dataset to get a 1.49 MB model with a 24.94 PSNR
score.
Related papers
- G-EvoNAS: Evolutionary Neural Architecture Search Based on Network
Growth [6.712149832731174]
This paper proposes a computationally efficient neural architecture evolutionary search framework based on network growth (G-EvoNAS)
The G-EvoNAS is tested on three commonly used image classification datasets, CIFAR10, CIFAR100, and ImageNet.
Experimental results demonstrate that G-EvoNAS can find a neural network architecture comparable to state-of-the-art designs in 0.2 GPU days.
arXiv Detail & Related papers (2024-03-05T05:44:38Z) - EAutoDet: Efficient Architecture Search for Object Detection [110.99532343155073]
EAutoDet framework can discover practical backbone and FPN architectures for object detection in 1.4 GPU-days.
We propose a kernel reusing technique by sharing the weights of candidate operations on one edge and consolidating them into one convolution.
In particular, the discovered architectures surpass state-of-the-art object detection NAS methods and achieve 40.1 mAP with 120 FPS and 49.2 mAP with 41.3 FPS on COCO test-dev set.
arXiv Detail & Related papers (2022-03-21T05:56:12Z) - Accelerating Training and Inference of Graph Neural Networks with Fast
Sampling and Pipelining [58.10436813430554]
Mini-batch training of graph neural networks (GNNs) requires a lot of computation and data movement.
We argue in favor of performing mini-batch training with neighborhood sampling in a distributed multi-GPU environment.
We present a sequence of improvements to mitigate these bottlenecks, including a performance-engineered neighborhood sampler.
We also conduct an empirical analysis that supports the use of sampling for inference, showing that test accuracies are not materially compromised.
arXiv Detail & Related papers (2021-10-16T02:41:35Z) - Combined Depth Space based Architecture Search For Person
Re-identification [70.86236888223569]
We aim to design a lightweight and suitable network for person re-identification (ReID)
We propose a novel search space called Combined Depth Space (CDS), based on which we search for an efficient network architecture, which we call CDNet.
We then propose a low-cost search strategy named the Top-k Sample Search strategy to make full use of the search space and avoid trapping in local optimal result.
arXiv Detail & Related papers (2021-04-09T02:40:01Z) - Fine-Grained Stochastic Architecture Search [6.277767522867666]
Fine-Grained Architecture Search (FiGS) is a differentiable search method that searches over a much larger set of candidate architectures.
FiGS simultaneously selects and modifies operators in the search space by applying a structured sparse regularization penalty.
We show results across 3 existing search spaces, matching or outperforming the original search algorithms.
arXiv Detail & Related papers (2020-06-17T01:04:14Z) - AlphaGAN: Fully Differentiable Architecture Search for Generative
Adversarial Networks [15.740179244963116]
Generative Adversarial Networks (GANs) are formulated as minimax game problems, whereby generators attempt to approach real data distributions by virtue of adversarial learning against discriminators.
In this work, we aim to boost model learning from the perspective of network architectures, by incorporating recent progress on automated architecture search into GANs.
We propose a fully differentiable search framework for generative adversarial networks, dubbed alphaGAN.
arXiv Detail & Related papers (2020-06-16T13:27:30Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z) - Real-Time Semantic Segmentation via Auto Depth, Downsampling Joint
Decision and Feature Aggregation [54.28963233377946]
We propose a joint search framework, called AutoRTNet, to automate the design of segmentation strategies.
Specifically, we propose hyper-cells to jointly decide the network depth and downsampling strategy, and an aggregation cell to achieve automatic multi-scale feature aggregation.
arXiv Detail & Related papers (2020-03-31T14:02:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.