WAS-VTON: Warping Architecture Search for Virtual Try-on Network
- URL: http://arxiv.org/abs/2108.00386v1
- Date: Sun, 1 Aug 2021 07:52:56 GMT
- Title: WAS-VTON: Warping Architecture Search for Virtual Try-on Network
- Authors: Zhenyu Xie, Xujie Zhang, Fuwei Zhao, Haoye Dong, Michael C.
Kampffmeyer, Haonan Yan, Xiaodan Liang
- Abstract summary: We introduce a NAS-Warping Module and elaborately design a bilevel hierarchical search space.
We learn a combination of repeatable warping cells and convolution operations specifically for the clothing-person alignment.
A NAS-Fusion Module is proposed to synthesize more natural final try-on results.
- Score: 57.52118202523266
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite recent progress on image-based virtual try-on, current methods are
constraint by shared warping networks and thus fail to synthesize natural
try-on results when faced with clothing categories that require different
warping operations. In this paper, we address this problem by finding clothing
category-specific warping networks for the virtual try-on task via Neural
Architecture Search (NAS). We introduce a NAS-Warping Module and elaborately
design a bilevel hierarchical search space to identify the optimal
network-level and operation-level flow estimation architecture. Given the
network-level search space, containing different numbers of warping blocks, and
the operation-level search space with different convolution operations, we
jointly learn a combination of repeatable warping cells and convolution
operations specifically for the clothing-person alignment. Moreover, a
NAS-Fusion Module is proposed to synthesize more natural final try-on results,
which is realized by leveraging particular skip connections to produce
better-fused features that are required for seamlessly fusing the warped
clothing and the unchanged person part. We adopt an efficient and stable
one-shot searching strategy to search the above two modules. Extensive
experiments demonstrate that our WAS-VTON significantly outperforms the
previous fixed-architecture try-on methods with more natural warping results
and virtual try-on results.
Related papers
- Neural Architecture Search by Learning a Hierarchical Search Space [7.154749344931972]
Monte-Carlo Tree Search (MCTS) is a powerful tool for many non-differentiable search related problems such as adversarial games.
In Neural Architecture Search (NAS), as only the final architecture matters, the visiting order of the branching can be optimized to improve learning.
We propose to learn the branching by hierarchical clustering of architectures based on their similarity.
arXiv Detail & Related papers (2025-03-27T00:20:13Z) - CSCO: Connectivity Search of Convolutional Operators [12.928148870505375]
We propose CSCO, a novel paradigm that fabricates effective connectivity of convolutional operators.
CSCO guides the exploration via a neural predictor as a surrogate of the ground-truth performance.
Results on ImageNet show 0.6% performance improvement over hand-crafted and NAS-crafted dense connectivity.
arXiv Detail & Related papers (2024-04-26T04:52:45Z) - HKNAS: Classification of Hyperspectral Imagery Based on Hyper Kernel
Neural Architecture Search [104.45426861115972]
We propose to directly generate structural parameters by utilizing the specifically designed hyper kernels.
We obtain three kinds of networks to separately conduct pixel-level or image-level classifications with 1-D or 3-D convolutions.
A series of experiments on six public datasets demonstrate that the proposed methods achieve state-of-the-art results.
arXiv Detail & Related papers (2023-04-23T17:27:40Z) - Pruning-as-Search: Efficient Neural Architecture Search via Channel
Pruning and Structural Reparameterization [50.50023451369742]
Pruning-as-Search (PaS) is an end-to-end channel pruning method to search out desired sub-network automatically and efficiently.
Our proposed architecture outperforms prior arts by around $1.0%$ top-1 accuracy on ImageNet-1000 classification task.
arXiv Detail & Related papers (2022-06-02T17:58:54Z) - Trilevel Neural Architecture Search for Efficient Single Image
Super-Resolution [127.92235484598811]
This paper proposes a trilevel neural architecture search (NAS) method for efficient single image super-resolution (SR)
For modeling the discrete search space, we apply a new continuous relaxation on the discrete search spaces to build a hierarchical mixture of network-path, cell-operations, and kernel-width.
An efficient search algorithm is proposed to perform optimization in a hierarchical supernet manner.
arXiv Detail & Related papers (2021-01-17T12:19:49Z) - Auto-Panoptic: Cooperative Multi-Component Architecture Search for
Panoptic Segmentation [144.50154657257605]
We propose an efficient framework to simultaneously search for all main components including backbone, segmentation branches, and feature fusion module.
Our searched architecture, namely Auto-Panoptic, achieves the new state-of-the-art on the challenging COCO and ADE20K benchmarks.
arXiv Detail & Related papers (2020-10-30T08:34:35Z) - Progressive Automatic Design of Search Space for One-Shot Neural
Architecture Search [15.017964136568061]
It has been observed that a model with higher one-shot model accuracy does not necessarily perform better when stand-alone trained.
We propose Progressive Automatic Design of search space, named PAD-NAS.
In this way, PAD-NAS can automatically design the operations for each layer and achieve a trade-off between search space quality and model diversity.
arXiv Detail & Related papers (2020-05-15T14:21:07Z) - NAS-Count: Counting-by-Density with Neural Architecture Search [74.92941571724525]
We automate the design of counting models with Neural Architecture Search (NAS)
We introduce an end-to-end searched encoder-decoder architecture, Automatic Multi-Scale Network (AMSNet)
arXiv Detail & Related papers (2020-02-29T09:18:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.