Learning High-Quality Initial Noise for Single-View Synthesis with Diffusion Models
- URL: http://arxiv.org/abs/2512.16219v1
- Date: Thu, 18 Dec 2025 06:08:21 GMT
- Title: Learning High-Quality Initial Noise for Single-View Synthesis with Diffusion Models
- Authors: Zhihao Zhang, Xuejun Yang, Weihua Liu, Mouquan Shen,
- Abstract summary: In diffusion models, certain high-quality initial noise patterns lead to better generation results than others.<n>We propose a learning framework based on an encoder-decoder network (EDN) that directly transforms random noise into high-quality noise.
- Score: 10.275373477634217
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Single-view novel view synthesis (NVS) models based on diffusion models have recently attracted increasing attention, as they can generate a series of novel view images from a single image prompt and camera pose information as conditions. It has been observed that in diffusion models, certain high-quality initial noise patterns lead to better generation results than others. However, there remains a lack of dedicated learning frameworks that enable NVS models to learn such high-quality noise. To obtain high-quality initial noise from random Gaussian noise, we make the following contributions. First, we design a discretized Euler inversion method to inject image semantic information into random noise, thereby constructing paired datasets of random and high-quality noise. Second, we propose a learning framework based on an encoder-decoder network (EDN) that directly transforms random noise into high-quality noise. Experiments demonstrate that the proposed EDN can be seamlessly plugged into various NVS models, such as SV3D and MV-Adapter, achieving significant performance improvements across multiple datasets. Code is available at: https://github.com/zhihao0512/EDN.
Related papers
- It's Never Too Late: Noise Optimization for Collapse Recovery in Trained Diffusion Models [80.53672733210111]
We show that a simple noise optimization objective can mitigate mode collapse while preserving the fidelity of the base model.<n>Our experiments demonstrate that noise optimization yields superior results in terms of generation quality and variety.
arXiv Detail & Related papers (2025-12-31T19:47:49Z) - GuidNoise: Single-Pair Guided Diffusion for Generalized Noise Synthesis [9.253859022117306]
Single-Pair Guided Diffusion for generalized noise synthesis GuidNoise.<n>GuidNoise uses a single noisy/clean pair as the guidance, often easily obtained by itself within a training set.<n>Uses a guidance-aware affine feature modification (GAFM) and a noise-aware refine loss to leverage the inherent potential of diffusion models.
arXiv Detail & Related papers (2025-12-04T05:00:00Z) - Revealing the Implicit Noise-based Imprint of Generative Models [71.94916898756684]
This paper presents a novel framework that leverages noise-based model-specific imprint for the detection task.<n>By aggregating imprints from various generative models, imprints of future models can be extrapolated to expand training data.<n>Our approach achieves state-of-the-art performance across three public benchmarks including GenImage, Synthbuster and Chameleon.
arXiv Detail & Related papers (2025-03-12T12:04:53Z) - One Noise to Rule Them All: Learning a Unified Model of Spatially-Varying Noise Patterns [33.293193191683145]
We present a single generative model which can learn to generate multiple types of noise as well as blend between them.
We also present an application of our model to improving inverse procedural material design.
arXiv Detail & Related papers (2024-04-25T02:23:11Z) - Blue noise for diffusion models [50.99852321110366]
We introduce a novel and general class of diffusion models taking correlated noise within and across images into account.
Our framework allows introducing correlation across images within a single mini-batch to improve gradient flow.
We perform both qualitative and quantitative evaluations on a variety of datasets using our method.
arXiv Detail & Related papers (2024-02-07T14:59:25Z) - DiffiT: Diffusion Vision Transformers for Image Generation [88.08529836125399]
Vision Transformer (ViT) has demonstrated strong modeling capabilities and scalability, especially for recognition tasks.
We study the effectiveness of ViTs in diffusion-based generative learning and propose a new model denoted as Diffusion Vision Transformers (DiffiT)
DiffiT is surprisingly effective in generating high-fidelity images with significantly better parameter efficiency.
arXiv Detail & Related papers (2023-12-04T18:57:01Z) - Towards General Low-Light Raw Noise Synthesis and Modeling [37.87312467017369]
We introduce a new perspective to synthesize the signal-independent noise by a generative model.
Specifically, we synthesize the signal-dependent and signal-independent noise in a physics- and learning-based manner.
In this way, our method can be considered as a general model, that is, it can simultaneously learn different noise characteristics for different ISO levels.
arXiv Detail & Related papers (2023-07-31T09:10:10Z) - Enhancing convolutional neural network generalizability via low-rank weight approximation [6.763245393373041]
Sufficient denoising is often an important first step for image processing.
Deep neural networks (DNNs) have been widely used for image denoising.
We introduce a new self-supervised framework for image denoising based on the Tucker low-rank tensor approximation.
arXiv Detail & Related papers (2022-09-26T14:11:05Z) - Noise2NoiseFlow: Realistic Camera Noise Modeling without Clean Images [35.29066692454865]
This paper proposes a framework for training a noise model and a denoiser simultaneously.
It relies on pairs of noisy images rather than noisy/clean paired image data.
The trained denoiser is shown to significantly improve upon both supervised and weakly supervised baseline denoising approaches.
arXiv Detail & Related papers (2022-06-02T15:31:40Z) - Unsupervised Single Image Super-resolution Under Complex Noise [60.566471567837574]
This paper proposes a model-based unsupervised SISR method to deal with the general SISR task with unknown degradations.
The proposed method can evidently surpass the current state of the art (SotA) method (about 1dB PSNR) not only with a slighter model (0.34M vs. 2.40M) but also faster speed.
arXiv Detail & Related papers (2021-07-02T11:55:40Z) - Neighbor2Neighbor: Self-Supervised Denoising from Single Noisy Images [98.82804259905478]
We present Neighbor2Neighbor to train an effective image denoising model with only noisy images.
In detail, input and target used to train a network are images sub-sampled from the same noisy image.
A denoising network is trained on sub-sampled training pairs generated in the first stage, with a proposed regularizer as additional loss for better performance.
arXiv Detail & Related papers (2021-01-08T02:03:25Z) - Knowledge Distillation in Iterative Generative Models for Improved
Sampling Speed [0.0]
Iterative generative models, such as noise conditional score networks, produce high quality samples by gradually denoising an initial noise vector.
We establish a novel connection between knowledge distillation and image generation with a technique that distills a multi-step denoising process into a single step.
Our Denoising Student generates high quality samples comparable to GANs on the CIFAR-10 and CelebA datasets, without adversarial training.
arXiv Detail & Related papers (2021-01-07T06:12:28Z) - Variational Denoising Network: Toward Blind Noise Modeling and Removal [59.36166491196973]
Blind image denoising is an important yet very challenging problem in computer vision.
We propose a new variational inference method, which integrates both noise estimation and image denoising.
arXiv Detail & Related papers (2019-08-29T15:54:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.