RapidAI4EO: Mono- and Multi-temporal Deep Learning models for Updating
the CORINE Land Cover Product
- URL: http://arxiv.org/abs/2210.14624v1
- Date: Wed, 26 Oct 2022 11:08:13 GMT
- Title: RapidAI4EO: Mono- and Multi-temporal Deep Learning models for Updating
the CORINE Land Cover Product
- Authors: Priyash Bhugra, Benjamin Bischke, Christoph Werner, Robert Syrnicki,
Carolin Packbier, Patrick Helber, Caglar Senaras, Akhil Singh Rana, Tim
Davis, Wanda De Keersmaecker, Daniele Zanaga, Annett Wania, Ruben Van De
Kerchove, Giovanni Marchisio
- Abstract summary: We evaluate the performance of multi-temporal (monthly time series) compared to mono-temporal (single time step) satellite images for multi-label classification.
We incorporated time-series images using a LSTM model to assess whether or not multi-temporal signals from satellites improves CLC classification.
- Score: 0.36265845593635804
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the remote sensing community, Land Use Land Cover (LULC) classification
with satellite imagery is a main focus of current research activities. Accurate
and appropriate LULC classification, however, continues to be a challenging
task. In this paper, we evaluate the performance of multi-temporal (monthly
time series) compared to mono-temporal (single time step) satellite images for
multi-label classification using supervised learning on the RapidAI4EO dataset.
As a first step, we trained our CNN model on images at a single time step for
multi-label classification, i.e. mono-temporal. We incorporated time-series
images using a LSTM model to assess whether or not multi-temporal signals from
satellites improves CLC classification. The results demonstrate an improvement
of approximately 0.89% in classifying satellite imagery on 15 classes using a
multi-temporal approach on monthly time series images compared to the
mono-temporal approach. Using features from multi-temporal or mono-temporal
images, this work is a step towards an efficient change detection and land
monitoring approach.
Related papers
- SpectralEarth: Training Hyperspectral Foundation Models at Scale [47.93167977587301]
We introduce SpectralEarth, a large-scale multi-temporal dataset designed to pretrain hyperspectral foundation models.
We pretrain a series of foundation models on SpectralEarth using state-of-the-art self-supervised learning (SSL) algorithms.
We construct four downstream datasets for land-cover and crop-type mapping, providing benchmarks for model evaluation.
arXiv Detail & Related papers (2024-08-15T22:55:59Z) - SatSynth: Augmenting Image-Mask Pairs through Diffusion Models for Aerial Semantic Segmentation [69.42764583465508]
We explore the potential of generative image diffusion to address the scarcity of annotated data in earth observation tasks.
To the best of our knowledge, we are the first to generate both images and corresponding masks for satellite segmentation.
arXiv Detail & Related papers (2024-03-25T10:30:22Z) - SkySense: A Multi-Modal Remote Sensing Foundation Model Towards Universal Interpretation for Earth Observation Imagery [35.550999964460466]
We present SkySense, a generic billion-scale model, pre-trained on a curated multi-modal Remote Sensing dataset with 21.5 million temporal sequences.
To our best knowledge, SkySense is the largest Multi-Modal to date, whose modules can be flexibly combined or used individually to accommodate various tasks.
arXiv Detail & Related papers (2023-12-15T09:57:21Z) - Unsupervised CD in satellite image time series by contrastive learning
and feature tracking [15.148034487267635]
We propose a two-stage approach to unsupervised change detection in satellite image time-series using contrastive learning with feature tracking.
By deriving pseudo labels from pre-trained models and using feature tracking to propagate them among the image time-series, we improve the consistency of our pseudo labels and address the challenges of seasonal changes in long-term remote sensing image time-series.
arXiv Detail & Related papers (2023-04-22T11:19:19Z) - Revisiting Temporal Modeling for CLIP-based Image-to-Video Knowledge
Transferring [82.84513669453744]
Image-text pretrained models, e.g., CLIP, have shown impressive general multi-modal knowledge learned from large-scale image-text data pairs.
We revisit temporal modeling in the context of image-to-video knowledge transferring.
We present a simple and effective temporal modeling mechanism extending CLIP model to diverse video tasks.
arXiv Detail & Related papers (2023-01-26T14:12:02Z) - ViTs for SITS: Vision Transformers for Satellite Image Time Series [52.012084080257544]
We introduce a fully-attentional model for general Satellite Image Time Series (SITS) processing based on the Vision Transformer (ViT)
TSViT splits a SITS record into non-overlapping patches in space and time which are tokenized and subsequently processed by a factorized temporo-spatial encoder.
arXiv Detail & Related papers (2023-01-12T11:33:07Z) - Learning to Exploit Temporal Structure for Biomedical Vision-Language
Processing [53.89917396428747]
Self-supervised learning in vision-language processing exploits semantic alignment between imaging and text modalities.
We explicitly account for prior images and reports when available during both training and fine-tuning.
Our approach, named BioViL-T, uses a CNN-Transformer hybrid multi-image encoder trained jointly with a text model.
arXiv Detail & Related papers (2023-01-11T16:35:33Z) - SatMAE: Pre-training Transformers for Temporal and Multi-Spectral
Satellite Imagery [74.82821342249039]
We present SatMAE, a pre-training framework for temporal or multi-spectral satellite imagery based on Masked Autoencoder (MAE)
To leverage temporal information, we include a temporal embedding along with independently masking image patches across time.
arXiv Detail & Related papers (2022-07-17T01:35:29Z) - Multi-Modal Temporal Attention Models for Crop Mapping from Satellite
Time Series [7.379078963413671]
Motivated by the recent success of temporal attention-based methods across multiple crop mapping tasks, we propose to investigate how these models can be adapted to operate on several modalities.
We implement and evaluate multiple fusion schemes, including a novel approach and simple adjustments to the training procedure.
We show that most fusion schemes have advantages and drawbacks, making them relevant for specific settings.
We then evaluate the benefit of multimodality across several tasks: parcel classification, pixel-based segmentation, and panoptic parcel segmentation.
arXiv Detail & Related papers (2021-12-14T17:05:55Z) - A Contrastive Learning Approach to Auroral Identification and
Classification [0.8399688944263843]
We present a novel application of unsupervised learning to the task of auroral image classification.
We modify and adapt the Simple framework for Contrastive Learning of Representations (SimCLR) algorithm to learn representations of auroral images.
Our approach exceeds an established threshold for operational purposes, demonstrating readiness for deployment and utilization.
arXiv Detail & Related papers (2021-09-28T17:51:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.