Texture synthesis via projection onto multiscale, multilayer statistics
- URL: http://arxiv.org/abs/2105.10825v1
- Date: Sat, 22 May 2021 23:32:34 GMT
- Title: Texture synthesis via projection onto multiscale, multilayer statistics
- Authors: Jieqian He and Matthew Hirn
- Abstract summary: We present a new model for texture synthesis based on a multiscale, multilayer feature extractor.
We explain the necessity of the different types of pre-defined wavelet filters used in our model and the advantages of multilayer structures for image synthesis.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We provide a new model for texture synthesis based on a multiscale,
multilayer feature extractor. Within the model, textures are represented by a
set of statistics computed from ReLU wavelet coefficients at different layers,
scales and orientations. A new image is synthesized by matching the target
statistics via an iterative projection algorithm. We explain the necessity of
the different types of pre-defined wavelet filters used in our model and the
advantages of multilayer structures for image synthesis. We demonstrate the
power of our model by generating samples of high quality textures and providing
insights into deep representations for texture images.
Related papers
- Infinite Texture: Text-guided High Resolution Diffusion Texture Synthesis [61.189479577198846]
We present Infinite Texture, a method for generating arbitrarily large texture images from a text prompt.
Our approach fine-tunes a diffusion model on a single texture, and learns to embed that statistical distribution in the output domain of the model.
At generation time, our fine-tuned diffusion model is used through a score aggregation strategy to generate output texture images of arbitrary resolution on a single GPU.
arXiv Detail & Related papers (2024-05-13T21:53:09Z) - SAMPLING: Scene-adaptive Hierarchical Multiplane Images Representation
for Novel View Synthesis from a Single Image [60.52991173059486]
We introduce SAMPLING, a Scene-adaptive Hierarchical Multiplane Images Representation for Novel View Synthesis from a Single Image.
Our method demonstrates considerable performance gains in large-scale unbounded outdoor scenes using a single image on the KITTI dataset.
arXiv Detail & Related papers (2023-09-12T15:33:09Z) - Learning in a Single Domain for Non-Stationary Multi-Texture Synthesis [9.213030142986417]
Non-stationary textures have large scale variance and can hardly be synthesized through one model.
We propose a multi-scale generator to capture structural patterns of various scales and effectively synthesize textures with a minor cost.
We present a category-specific training strategy to focus on learning texture pattern of a specific domain.
arXiv Detail & Related papers (2023-05-10T14:32:21Z) - DiVAE: Photorealistic Images Synthesis with Denoising Diffusion Decoder [73.1010640692609]
We propose a VQ-VAE architecture model with a diffusion decoder (DiVAE) to work as the reconstructing component in image synthesis.
Our model achieves state-of-the-art results and generates more photorealistic images specifically.
arXiv Detail & Related papers (2022-06-01T10:39:12Z) - RTMV: A Ray-Traced Multi-View Synthetic Dataset for Novel View Synthesis [104.53930611219654]
We present a large-scale synthetic dataset for novel view synthesis consisting of 300k images rendered from nearly 2000 complex scenes.
The dataset is orders of magnitude larger than existing synthetic datasets for novel view synthesis.
Using 4 distinct sources of high-quality 3D meshes, the scenes of our dataset exhibit challenging variations in camera views, lighting, shape, materials, and textures.
arXiv Detail & Related papers (2022-05-14T13:15:32Z) - Multiscale Analysis for Improving Texture Classification [62.226224120400026]
This paper employs the Gaussian-Laplacian pyramid to treat different spatial frequency bands of a texture separately.
We aggregate features extracted from gray and color texture images using bio-inspired texture descriptors, information-theoretic measures, gray-level co-occurrence matrix features, and Haralick statistical features into a single feature vector.
arXiv Detail & Related papers (2022-04-21T01:32:22Z) - Generalized Rectifier Wavelet Covariance Models For Texture Synthesis [2.585403833659771]
State-of-the-art maximum entropy models for texture synthesis are built from statistics relying on image representations defined by convolutional neural networks (CNN)
We propose a family of statistics built upon non-linear wavelet based representations, that can be viewed as a particular instance of a one-layer CNN, using a generalized non-linearity.
These statistics significantly improve the visual quality of previous classical wavelet-based models, and allow one to produce syntheses of similar quality to state-of-the-art models, on both grayscale and color textures.
arXiv Detail & Related papers (2022-03-14T17:07:40Z) - A Generative Model for Texture Synthesis based on Optimal Transport
between Feature Distributions [8.102785819558978]
We show how to use our framework to learn a feed-forward neural network that can synthesize on-the-fly new textures of arbitrary size.
We show how to use our framework to learn a feed-forward neural network that can synthesize on-the-fly new textures of arbitrary size in a very fast manner.
arXiv Detail & Related papers (2020-06-19T13:32:55Z) - Region-adaptive Texture Enhancement for Detailed Person Image Synthesis [86.69934638569815]
RATE-Net is a novel framework for synthesizing person images with sharp texture details.
The proposed framework leverages an additional texture enhancing module to extract appearance information from the source image.
Experiments conducted on DeepFashion benchmark dataset have demonstrated the superiority of our framework compared with existing networks.
arXiv Detail & Related papers (2020-05-26T02:33:21Z) - Semantic Pyramid for Image Generation [41.85213024720986]
We present a novel GAN-based model that utilizes the space of deep features learned by a pre-trained classification model.
Inspired by classical image pyramid representations, we construct our model as a Semantic Generation Pyramid.
arXiv Detail & Related papers (2020-03-13T12:23:37Z) - Contextual Encoder-Decoder Network for Visual Saliency Prediction [42.047816176307066]
We propose an approach based on a convolutional neural network pre-trained on a large-scale image classification task.
We combine the resulting representations with global scene information for accurately predicting visual saliency.
Compared to state of the art approaches, the network is based on a lightweight image classification backbone.
arXiv Detail & Related papers (2019-02-18T16:15:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.