Cross-modal Ship Re-Identification via Optical and SAR Imagery: A Novel Dataset and Method
- URL: http://arxiv.org/abs/2506.22027v3
- Date: Wed, 16 Jul 2025 02:40:11 GMT
- Title: Cross-modal Ship Re-Identification via Optical and SAR Imagery: A Novel Dataset and Method
- Authors: Han Wang, Shengyang Li, Jian Yang, Yuxuan Liu, Yixuan Lv, Zhuang Zhou,
- Abstract summary: Continuous maritime ship tracking is crucial for applications such as maritime search and rescue, law enforcement, and shipping analysis.<n>Most current ship tracking methods rely on geostationary satellites or video satellites.<n>To address these limitations, we present the Hybrid Optical and Synthetic Aperture Radar (SAR) Ship Re-Identification dataset.
- Score: 16.8202125093375
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Detecting and tracking ground objects using earth observation imagery remains a significant challenge in the field of remote sensing. Continuous maritime ship tracking is crucial for applications such as maritime search and rescue, law enforcement, and shipping analysis. However, most current ship tracking methods rely on geostationary satellites or video satellites. The former offer low resolution and are susceptible to weather conditions, while the latter have short filming durations and limited coverage areas, making them less suitable for the real-world requirements of ship tracking. To address these limitations, we present the Hybrid Optical and Synthetic Aperture Radar (SAR) Ship Re-Identification Dataset (HOSS ReID dataset), designed to evaluate the effectiveness of ship tracking using low-Earth orbit constellations of optical and SAR sensors. This approach ensures shorter re-imaging cycles and enables all-weather tracking. HOSS ReID dataset includes images of the same ship captured over extended periods under diverse conditions, using different satellites of different modalities at varying times and angles. Furthermore, we propose a baseline method for cross-modal ship re-identification, TransOSS, which is built on the Vision Transformer architecture. It refines the patch embedding structure to better accommodate cross-modal tasks, incorporates additional embeddings to introduce more reference information, and employs contrastive learning to pre-train on large-scale optical-SAR image pairs, ensuring the model's ability to extract modality-invariant features. Our dataset and baseline method are publicly available on https://github.com/Alioth2000/Hoss-ReID.
Related papers
- SMART-Ship: A Comprehensive Synchronized Multi-modal Aligned Remote Sensing Targets Dataset and Benchmark for Berthed Ships Analysis [12.87083600993665]
This dataset consists of 1092 multi-modal image sets, covering 38,838 ships.<n>Each image set is acquired within one week and registered to ensuretemporal consistency.<n>We define benchmarks on five fundamental tasks and compare methods across the dataset.
arXiv Detail & Related papers (2025-08-04T13:09:58Z) - NOVA: Navigation via Object-Centric Visual Autonomy for High-Speed Target Tracking in Unstructured GPS-Denied Environments [56.35569661650558]
We introduce NOVA, a fully onboard, object-centric framework that enables robust target tracking and collision-aware navigation.<n>Rather than constructing a global map, NOVA formulates perception, estimation, and control entirely in the target's reference frame.<n>We validate NOVA across challenging real-world scenarios, including urban mazes, forest trails, and repeated transitions through buildings with intermittent GPS loss.
arXiv Detail & Related papers (2025-06-23T14:28:30Z) - MOANA: Multi-Radar Dataset for Maritime Odometry and Autonomous Navigation Application [10.093577014949398]
This dataset integrates short-range LiDAR data, medium-range W-band radar data, and long-range X-band radar data into a unified framework.<n>The dataset comprises seven sequences collected from diverse regions with varying levels of blnavigation algorithm estimation difficulty.
arXiv Detail & Related papers (2024-12-05T05:40:40Z) - Mixing Data-driven and Geometric Models for Satellite Docking Port State Estimation using an RGB or Event Camera [4.9788231201543]
This work focuses on satellite-agnostic operations using the recently released Lockheed Martin Mission Augmentation Port (LM-MAP) as the target.
We present a pipeline for automated satellite docking port detection and state estimation using monocular vision data from standard RGB sensing or an event camera.
arXiv Detail & Related papers (2024-09-23T22:28:09Z) - On the Generation of a Synthetic Event-Based Vision Dataset for
Navigation and Landing [69.34740063574921]
This paper presents a methodology for generating event-based vision datasets from optimal landing trajectories.
We construct sequences of photorealistic images of the lunar surface with the Planet and Asteroid Natural Scene Generation Utility.
We demonstrate that the pipeline can generate realistic event-based representations of surface features by constructing a dataset of 500 trajectories.
arXiv Detail & Related papers (2023-08-01T09:14:20Z) - xView3-SAR: Detecting Dark Fishing Activity Using Synthetic Aperture
Radar Imagery [52.67592123500567]
Unsustainable fishing practices worldwide pose a major threat to marine resources and ecosystems.
It is now possible to automate detection of dark vessels day or night, under all-weather conditions.
xView3-SAR consists of nearly 1,000 analysis-ready SAR images from the Sentinel-1 mission.
arXiv Detail & Related papers (2022-06-02T06:53:45Z) - Deep Learning for Real Time Satellite Pose Estimation on Low Power Edge
TPU [58.720142291102135]
In this paper we propose a pose estimation software exploiting neural network architectures.
We show how low power machine learning accelerators could enable Artificial Intelligence exploitation in space.
arXiv Detail & Related papers (2022-04-07T08:53:18Z) - Towards Robust Monocular Visual Odometry for Flying Robots on Planetary
Missions [49.79068659889639]
Ingenuity, that just landed on Mars, will mark the beginning of a new era of exploration unhindered by traversability.
We present an advanced robust monocular odometry algorithm that uses efficient optical flow tracking.
We also present a novel approach to estimate the current risk of scale drift based on a principal component analysis of the relative translation information matrix.
arXiv Detail & Related papers (2021-09-12T12:52:20Z) - FGSD: A Dataset for Fine-Grained Ship Detection in High Resolution
Satellite Images [22.883300168530035]
Ship detection using high-resolution remote sensing images is an important task, which contribute to sea surface regulation.
To promote the research of ship detection, we introduced a new fine-grained ship detection datasets, which is named as FGSD.
The dataset collects high-resolution remote sensing images that containing ship samples from multiple large ports around the world.
arXiv Detail & Related papers (2020-03-15T13:54:20Z) - Drone-based RGB-Infrared Cross-Modality Vehicle Detection via
Uncertainty-Aware Learning [59.19469551774703]
Drone-based vehicle detection aims at finding the vehicle locations and categories in an aerial image.
We construct a large-scale drone-based RGB-Infrared vehicle detection dataset, termed DroneVehicle.
Our DroneVehicle collects 28, 439 RGB-Infrared image pairs, covering urban roads, residential areas, parking lots, and other scenarios from day to night.
arXiv Detail & Related papers (2020-03-05T05:29:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.