CloudMatch: Weak-to-Strong Consistency Learning for Semi-Supervised Cloud Detection
- URL: http://arxiv.org/abs/2601.03528v1
- Date: Wed, 07 Jan 2026 02:31:17 GMT
- Title: CloudMatch: Weak-to-Strong Consistency Learning for Semi-Supervised Cloud Detection
- Authors: Jiayi Zhao, Changlu Chen, Jingsheng Li, Tianxiang Xue, Kun Zhan,
- Abstract summary: CloudMatch is a semi-supervised framework that effectively leverages unlabeled remote sensing imagery.<n>Our key insight is that enforcing prediction consistency across diversely augmented views, incorporating both inter-scene and intra-scene mixing.<n>Experiments show that CloudMatch achieves good performance, demonstrating its capability to utilize unlabeled data efficiently.
- Score: 10.250611987029254
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Due to the high cost of annotating accurate pixel-level labels, semi-supervised learning has emerged as a promising approach for cloud detection. In this paper, we propose CloudMatch, a semi-supervised framework that effectively leverages unlabeled remote sensing imagery through view-consistency learning combined with scene-mixing augmentations. An observation behind CloudMatch is that cloud patterns exhibit structural diversity and contextual variability across different scenes and within the same scene category. Our key insight is that enforcing prediction consistency across diversely augmented views, incorporating both inter-scene and intra-scene mixing, enables the model to capture the structural diversity and contextual richness of cloud patterns. Specifically, CloudMatch generates one weakly augmented view along with two complementary strongly augmented views for each unlabeled image: one integrates inter-scene patches to simulate contextual variety, while the other employs intra-scene mixing to preserve semantic coherence. This approach guides pseudolabel generation and enhances generalization. Extensive experiments show that CloudMatch achieves good performance, demonstrating its capability to utilize unlabeled data efficiently and advance semi-supervised cloud detection.
Related papers
- Weakly Supervised Cloud Detection Combining Spectral Features and Multi-Scale Deep Network [12.520904004953344]
We propose a weakly supervised cloud detection method that combines spectral features and multi-scale scene-level deep network (SpecMCD) to obtain highly accurate pixel-level cloud masks.<n>The F1-score of the proposed SpecMCD method shows an improvement of over 7.82%, highlighting the superiority and potential of the SpecMCD method for cloud detection.
arXiv Detail & Related papers (2025-10-01T08:32:49Z) - Masked Clustering Prediction for Unsupervised Point Cloud Pre-training [61.11226004056774]
MaskClu is a novel unsupervised pre-training method for ViTs on 3D point clouds.<n>It integrates masked point modeling with clustering-based learning.
arXiv Detail & Related papers (2025-08-12T12:58:44Z) - Towards Fusing Point Cloud and Visual Representations for Imitation Learning [57.886331184389604]
We propose FPV-Net, a novel imitation learning method that effectively combines the strengths of both point cloud and RGB modalities.<n>Our method conditions the point-cloud encoder on global and local image tokens using adaptive layer norm conditioning.
arXiv Detail & Related papers (2025-02-17T20:46:54Z) - HVDistill: Transferring Knowledge from Images to Point Clouds via Unsupervised Hybrid-View Distillation [106.09886920774002]
We present a hybrid-view-based knowledge distillation framework, termed HVDistill, to guide the feature learning of a point cloud neural network.
Our method achieves consistent improvements over the baseline trained from scratch and significantly out- performs the existing schemes.
arXiv Detail & Related papers (2024-03-18T14:18:08Z) - See More and Know More: Zero-shot Point Cloud Segmentation via
Multi-modal Visual Data [22.53879737713057]
Zero-shot point cloud segmentation aims to make deep models capable of recognizing novel objects in point cloud that are unseen in the training phase.
We propose a novel multi-modal zero-shot learning method to better utilize the complementary information of point clouds and images for more accurate visual-semantic alignment.
arXiv Detail & Related papers (2023-07-20T11:32:51Z) - Contrastive Predictive Autoencoders for Dynamic Point Cloud
Self-Supervised Learning [26.773995001469505]
We design point cloud sequence based Contrastive Prediction and Reconstruction (CPR), to collaboratively learn more comprehensive representations.
We conduct experiments on four point cloud sequence benchmarks, and report the results under multiple experimental settings.
arXiv Detail & Related papers (2023-05-22T12:09:51Z) - EPCL: Frozen CLIP Transformer is An Efficient Point Cloud Encoder [60.52613206271329]
This paper introduces textbfEfficient textbfPoint textbfCloud textbfLearning (EPCL) for training high-quality point cloud models with a frozen CLIP transformer.
Our EPCL connects the 2D and 3D modalities by semantically aligning the image features and point cloud features without paired 2D-3D data.
arXiv Detail & Related papers (2022-12-08T06:27:11Z) - Data Augmentation-free Unsupervised Learning for 3D Point Cloud
Understanding [61.30276576646909]
We propose an augmentation-free unsupervised approach for point clouds to learn transferable point-level features via soft clustering, named SoftClu.
We exploit the affiliation of points to their clusters as a proxy to enable self-training through a pseudo-label prediction task.
arXiv Detail & Related papers (2022-10-06T10:18:16Z) - ImLoveNet: Misaligned Image-supported Registration Network for
Low-overlap Point Cloud Pairs [14.377604289952188]
Low-overlap regions between paired point clouds make the captured features very low-confidence.
We propose a misaligned image supported registration network for low-overlap point cloud pairs, dubbed ImLoveNet.
arXiv Detail & Related papers (2022-07-02T13:17:34Z) - Unsupervised Representation Learning for 3D Point Cloud Data [66.92077180228634]
We propose a simple yet effective approach for unsupervised point cloud learning.
In particular, we identify a very useful transformation which generates a good contrastive version of an original point cloud.
We conduct experiments on three downstream tasks which are 3D object classification, shape part segmentation and scene segmentation.
arXiv Detail & Related papers (2021-10-13T10:52:45Z) - Single Image Cloud Detection via Multi-Image Fusion [23.641624507709274]
A primary challenge in developing algorithms is the cost of collecting annotated training data.
We demonstrate how recent advances in multi-image fusion can be leveraged to bootstrap single image cloud detection.
We collect a large dataset of Sentinel-2 images along with a per-pixel semantic labelling for land cover.
arXiv Detail & Related papers (2020-07-29T22:52:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.