Texture Reformer: Towards Fast and Universal Interactive Texture
Transfer
- URL: http://arxiv.org/abs/2112.02788v1
- Date: Mon, 6 Dec 2021 05:20:43 GMT
- Title: Texture Reformer: Towards Fast and Universal Interactive Texture
Transfer
- Authors: Zhizhong Wang, Lei Zhao, Haibo Chen, Ailin Li, Zhiwen Zuo, Wei Xing,
Dongming Lu
- Abstract summary: texture reformer is a neural-based framework for interactive texture transfer with user-specified guidance.
We introduce a novel learning-free view-specific texture reformation (VSTR) operation with a new semantic map guidance strategy.
The experimental results on a variety of application scenarios demonstrate the effectiveness and superiority of our framework.
- Score: 16.41438144343516
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present the texture reformer, a fast and universal
neural-based framework for interactive texture transfer with user-specified
guidance. The challenges lie in three aspects: 1) the diversity of tasks, 2)
the simplicity of guidance maps, and 3) the execution efficiency. To address
these challenges, our key idea is to use a novel feed-forward multi-view and
multi-stage synthesis procedure consisting of I) a global view structure
alignment stage, II) a local view texture refinement stage, and III) a holistic
effect enhancement stage to synthesize high-quality results with coherent
structures and fine texture details in a coarse-to-fine fashion. In addition,
we also introduce a novel learning-free view-specific texture reformation
(VSTR) operation with a new semantic map guidance strategy to achieve more
accurate semantic-guided and structure-preserved texture transfer. The
experimental results on a variety of application scenarios demonstrate the
effectiveness and superiority of our framework. And compared with the
state-of-the-art interactive texture transfer algorithms, it not only achieves
higher quality results but, more remarkably, also is 2-5 orders of magnitude
faster. Code is available at https://github.com/EndyWon/Texture-Reformer.
Related papers
- Meta 3D TextureGen: Fast and Consistent Texture Generation for 3D Objects [54.80813150893719]
We introduce Meta 3D TextureGen: a new feedforward method comprised of two sequential networks aimed at generating high-quality textures in less than 20 seconds.
Our method state-of-the-art results in quality and speed by conditioning a text-to-image model on 3D semantics in 2D space and fusing them into a complete and high-resolution UV texture map.
In addition, we introduce a texture enhancement network that is capable of up-scaling any texture by an arbitrary ratio, producing 4k pixel resolution textures.
arXiv Detail & Related papers (2024-07-02T17:04:34Z) - InTeX: Interactive Text-to-texture Synthesis via Unified Depth-aware Inpainting [46.330305910974246]
We introduce InteX, a novel framework for interactive text-to-texture synthesis.
InteX includes a user-friendly interface that facilitates interaction and control throughout the synthesis process.
We develop a depth-aware inpainting model that integrates depth information with inpainting cues, effectively mitigating 3D inconsistencies.
arXiv Detail & Related papers (2024-03-18T15:31:57Z) - TextureDreamer: Image-guided Texture Synthesis through Geometry-aware
Diffusion [64.49276500129092]
TextureDreamer is an image-guided texture synthesis method.
It can transfer relightable textures from a small number of input images to target 3D shapes across arbitrary categories.
arXiv Detail & Related papers (2024-01-17T18:55:49Z) - TextFormer: A Query-based End-to-End Text Spotter with Mixed Supervision [61.186488081379]
We propose TextFormer, a query-based end-to-end text spotter with Transformer architecture.
TextFormer builds upon an image encoder and a text decoder to learn a joint semantic understanding for multi-task modeling.
It allows for mutual training and optimization of classification, segmentation, and recognition branches, resulting in deeper feature sharing.
arXiv Detail & Related papers (2023-06-06T03:37:41Z) - Towards Universal Texture Synthesis by Combining Texton Broadcasting
with Noise Injection in StyleGAN-2 [11.67779950826776]
We present a new approach for universal texture synthesis incorporating by a multi-scale texton broadcasting module in the StyleGAN-2 framework.
The texton broadcasting module introduces an inductive bias, enabling generation of broader range of textures from those with regular structures to completely ones.
arXiv Detail & Related papers (2022-03-08T17:44:35Z) - Paying U-Attention to Textures: Multi-Stage Hourglass Vision Transformer for Universal Texture Synthesis [2.8998926117101367]
We present a novel U-Attention vision Transformer for universal texture synthesis.
We exploit the natural long-range dependencies enabled by the attention mechanism to allow our approach to synthesize diverse textures.
We propose a hierarchical hourglass backbone that attends to the global structure and performs patch mapping at varying scales.
arXiv Detail & Related papers (2022-02-23T18:58:56Z) - Image Inpainting via Conditional Texture and Structure Dual Generation [26.97159780261334]
We propose a novel two-stream network for image inpainting, which models the structure-constrained texture synthesis and texture-guided structure reconstruction.
To enhance the global consistency, a Bi-directional Gated Feature Fusion (Bi-GFF) module is designed to exchange and combine the structure and texture information.
Experiments on the CelebA, Paris StreetView and Places2 datasets demonstrate the superiority of the proposed method.
arXiv Detail & Related papers (2021-08-22T15:44:37Z) - DF-GAN: A Simple and Effective Baseline for Text-to-Image Synthesis [80.54273334640285]
We propose a novel one-stage text-to-image backbone that directly synthesizes high-resolution images without entanglements between different generators.
We also propose a novel Target-Aware Discriminator composed of Matching-Aware Gradient Penalty and One-Way Output.
Compared with current state-of-the-art methods, our proposed DF-GAN is simpler but more efficient to synthesize realistic and text-matching images.
arXiv Detail & Related papers (2020-08-13T12:51:17Z) - Region-adaptive Texture Enhancement for Detailed Person Image Synthesis [86.69934638569815]
RATE-Net is a novel framework for synthesizing person images with sharp texture details.
The proposed framework leverages an additional texture enhancing module to extract appearance information from the source image.
Experiments conducted on DeepFashion benchmark dataset have demonstrated the superiority of our framework compared with existing networks.
arXiv Detail & Related papers (2020-05-26T02:33:21Z) - Towards Analysis-friendly Face Representation with Scalable Feature and
Texture Compression [113.30411004622508]
We show that a universal and collaborative visual information representation can be achieved in a hierarchical way.
Based on the strong generative capability of deep neural networks, the gap between the base feature layer and enhancement layer is further filled with the feature level texture reconstruction.
To improve the efficiency of the proposed framework, the base layer neural network is trained in a multi-task manner.
arXiv Detail & Related papers (2020-04-21T14:32:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.