SADER: Structure-Aware Diffusion Framework with DEterministic Resampling for Multi-Temporal Remote Sensing Cloud Removal
- URL: http://arxiv.org/abs/2602.00536v1
- Date: Sat, 31 Jan 2026 06:06:18 GMT
- Title: SADER: Structure-Aware Diffusion Framework with DEterministic Resampling for Multi-Temporal Remote Sensing Cloud Removal
- Authors: Yifan Zhang, Qian Chen, Yi Liu, Wengen Li, Jihong Guan,
- Abstract summary: Diffusion-based models have emerged as a dominant paradigm for remote sensing cloud removal.<n>We propose SADER, a structure-aware diffusion framework for multi-temporal remote sensing cloud removal.<n>Experiments on multiple multi-temporal datasets demonstrate that SADER consistently outperforms state-of-the-art cloud removal methods.
- Score: 24.84158171817714
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cloud contamination severely degrades the usability of remote sensing imagery and poses a fundamental challenge for downstream Earth observation tasks. Recently, diffusion-based models have emerged as a dominant paradigm for remote sensing cloud removal due to their strong generative capability and stable optimization. However, existing diffusion-based approaches often suffer from limited sampling efficiency and insufficient exploitation of structural and temporal priors in multi-temporal remote sensing scenarios. In this work, we propose SADER, a structure-aware diffusion framework for multi-temporal remote sensing cloud removal. SADER first develops a scalable Multi-Temporal Conditional Diffusion Network (MTCDN) to fully capture multi-temporal and multimodal correlations via temporal fusion and hybrid attention. Then, a cloud-aware attention loss is introduced to emphasize cloud-dominated regions by accounting for cloud thickness and brightness discrepancies. In addition, a deterministic resampling strategy is designed for continuous diffusion models to iteratively refine samples under fixed sampling steps by replacing outliers through guided correction. Extensive experiments on multiple multi-temporal datasets demonstrate that SADER consistently outperforms state-of-the-art cloud removal methods across all evaluation metrics. The code of SADER is publicly available at https://github.com/zyfzs0/SADER.
Related papers
- Unleashing Temporal Capacity of Spiking Neural Networks through Spatiotemporal Separation [67.69345363409835]
Spiking Neural Networks (SNNs) are considered naturally suited for temporal processing, with membrane potential propagation widely regarded as the core temporal modeling mechanism.<n>We design Non-Stateful (NS) models progressively removing membrane propagation to its stage-wise role. Experiments reveal a counterintuitive phenomenon: moderate removal in shallow layers improves performance, while excessive removal causes collapse.
arXiv Detail & Related papers (2025-12-05T07:05:53Z) - USF-Net: A Unified Spatiotemporal Fusion Network for Ground-Based Remote Sensing Cloud Image Sequence Extrapolation [7.868367798549883]
Ground-based remote sensing cloud image sequence extrapolation is a key research area in the development of photovoltaic power systems.<n>We propose USF-Net, a Unified Stemporal Fusion that integrates adaptive large- kernel convolutions and a low-complexity attention mechanism.<n>As a key contribution, we also introduce and release the ASI-CIS dataset.
arXiv Detail & Related papers (2025-11-12T06:54:40Z) - RainDiff: End-to-end Precipitation Nowcasting Via Token-wise Attention Diffusion [64.49056527678606]
We propose a Token-wise Attention integrated into not only the U-Net diffusion model but also the radar-temporal encoder.<n>Unlike prior approaches, our method integrates attention into the architecture without incurring the high resource cost typical of pixel-space diffusion.<n>Our experiments and evaluations demonstrate that the proposed method significantly outperforms state-of-the-art approaches, robustness local fidelity, generalization, and superior in complex precipitation forecasting scenarios.
arXiv Detail & Related papers (2025-10-16T17:59:13Z) - EC-Diff: Fast and High-Quality Edge-Cloud Collaborative Inference for Diffusion Models [57.059991285047296]
hybrid edge-cloud collaborative framework was recently proposed to realize fast inference and high-quality generation.<n>Excessive cloud denoising prolongs inference time, while insufficient steps cause semantic ambiguity, leading to inconsistency in edge model output.<n>We propose EC-Diff that accelerates cloud inference through gradient-based noise estimation.<n>Our method significantly enhances generation quality compared to edge inference, while achieving up to an average $2times$ speedup in inference compared to cloud inference.
arXiv Detail & Related papers (2025-07-16T07:23:14Z) - When Cloud Removal Meets Diffusion Model in Remote Sensing [4.824120664293887]
We propose DC4CR (Diffusion Control for Cloud Removal), a novel framework for cloud removal in remote sensing imagery.<n>Our method introduces prompt-driven control, allowing selective removal of thin and thick clouds without relying on pre-generated cloud masks.<n>Experiments on the RICE and CUHK-CR datasets demonstrate state-of-the-art performance, achieving superior cloud removal across diverse conditions.
arXiv Detail & Related papers (2025-04-21T00:56:57Z) - Multimodal Diffusion Bridge with Attention-Based SAR Fusion for Satellite Image Cloud Removal [15.014162404892557]
Diffusion Bridges for Cloud Removal, DB-CR, bridges between cloudy and cloud-free image distributions.<n>DB-CR achieves high-fidelity results while being computationally efficient.
arXiv Detail & Related papers (2025-04-04T17:25:49Z) - Point Cloud Resampling with Learnable Heat Diffusion [58.050130177241186]
We propose a learnable heat diffusion framework for point cloud resampling.
Unlike previous diffusion models with a fixed prior, the adaptive conditional prior selectively preserves geometric features of the point cloud.
arXiv Detail & Related papers (2024-11-21T13:44:18Z) - Improving Vector-Quantized Image Modeling with Latent Consistency-Matching Diffusion [55.185588994883226]
We introduce VQ-LCMD, a continuous-space latent diffusion framework within the embedding space that stabilizes training.<n>VQ-LCMD uses a novel training objective combining the joint embedding-diffusion variational lower bound with a consistency-matching (CM) loss.<n>Experiments show that the proposed VQ-LCMD yields superior results on FFHQ, LSUN Churches, and LSUN Bedrooms compared to discrete-state latent diffusion models.
arXiv Detail & Related papers (2024-10-18T09:12:33Z) - Multi-scale Restoration of Missing Data in Optical Time-series Images with Masked Spatial-Temporal Attention Network [0.6675733925327885]
Existing methods for imputing missing values in remote sensing images fail to fully exploit auxiliary information.
This paper proposes a deep learning-based novel approach called MS2 for reconstructing time-series remote sensing images.
arXiv Detail & Related papers (2024-06-19T09:05:05Z) - IDF-CR: Iterative Diffusion Process for Divide-and-Conquer Cloud Removal in Remote-sensing Images [55.40601468843028]
We present an iterative diffusion process for cloud removal (IDF-CR)
IDF-CR is divided into two-stage models that address pixel space and latent space.
In the latent space stage, the diffusion model transforms low-quality cloud removal into high-quality clean output.
arXiv Detail & Related papers (2024-03-18T15:23:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.