Hierarchical Paired Channel Fusion Network for Street Scene Change
Detection
- URL: http://arxiv.org/abs/2010.09925v1
- Date: Mon, 19 Oct 2020 23:51:28 GMT
- Title: Hierarchical Paired Channel Fusion Network for Street Scene Change
Detection
- Authors: Yinjie Lei and Duo Peng and Pingping Zhang and Qiuhong Ke and Haifeng
Li
- Abstract summary: Street Scene Change Detection (SSCD) aims to locate the changed regions between a given street-view image pair captured at different times.
We propose a novel Hierarchical Paired Channel Fusion Network ( HPCFNet) to improve the accuracy of the corresponding change maps.
Our framework achieves a novel approach to adapt to the scale and location diversities of the scene change regions.
- Score: 41.934290847053695
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Street Scene Change Detection (SSCD) aims to locate the changed regions
between a given street-view image pair captured at different times, which is an
important yet challenging task in the computer vision community. The intuitive
way to solve the SSCD task is to fuse the extracted image feature pairs, and
then directly measure the dissimilarity parts for producing a change map.
Therefore, the key for the SSCD task is to design an effective feature fusion
method that can improve the accuracy of the corresponding change maps. To this
end, we present a novel Hierarchical Paired Channel Fusion Network (HPCFNet),
which utilizes the adaptive fusion of paired feature channels. Specifically,
the features of a given image pair are jointly extracted by a Siamese
Convolutional Neural Network (SCNN) and hierarchically combined by exploring
the fusion of channel pairs at multiple feature levels. In addition, based on
the observation that the distribution of scene changes is diverse, we further
propose a Multi-Part Feature Learning (MPFL) strategy to detect diverse
changes. Based on the MPFL strategy, our framework achieves a novel approach to
adapt to the scale and location diversities of the scene change regions.
Extensive experiments on three public datasets (i.e., PCD, VL-CMU-CD and
CDnet2014) demonstrate that the proposed framework achieves superior
performance which outperforms other state-of-the-art methods with a
considerable margin.
Related papers
- Robust Scene Change Detection Using Visual Foundation Models and Cross-Attention Mechanisms [27.882122236282054]
We present a novel method for scene change detection that leverages the robust feature extraction capabilities of a visual foundational model, DINOv2.
We evaluate our approach on two benchmark datasets, VL-CMU-CD and PSCD, along with their viewpoint-varied versions.
Our experiments demonstrate significant improvements in F1-score, particularly in scenarios involving geometric changes between image pairs.
arXiv Detail & Related papers (2024-09-25T11:55:27Z) - Cross-Domain Separable Translation Network for Multimodal Image Change Detection [11.25422609271201]
multimodal change detection (MCD) is particularly critical in the remote sensing community.
This paper focuses on addressing the challenges of MCD, especially the difficulty in comparing images from different sensors.
A novel unsupervised cross-domain separable translation network (CSTN) is proposed to overcome these limitations.
arXiv Detail & Related papers (2024-07-23T03:56:02Z) - ChangeBind: A Hybrid Change Encoder for Remote Sensing Change Detection [16.62779899494721]
Change detection (CD) is a fundamental task in remote sensing (RS) which aims to detect the semantic changes between the same geographical regions at different time stamps.
We propose an effective Siamese-based framework to encode the semantic changes occurring in the bi-temporal RS images.
arXiv Detail & Related papers (2024-04-26T17:47:14Z) - TransY-Net:Learning Fully Transformer Networks for Change Detection of
Remote Sensing Images [64.63004710817239]
We propose a novel Transformer-based learning framework named TransY-Net for remote sensing image CD.
It improves the feature extraction from a global view and combines multi-level visual features in a pyramid manner.
Our proposed method achieves a new state-of-the-art performance on four optical and two SAR image CD benchmarks.
arXiv Detail & Related papers (2023-10-22T07:42:19Z) - dual unet:a novel siamese network for change detection with cascade
differential fusion [4.651756476458979]
We propose a novel Siamese neural network for change detection task, namely Dual-UNet.
In contrast to previous individually encoded the bitemporal images, we design an encoder differential-attention module to focus on the spatial difference relationships of pixels.
Experiments demonstrate that the proposed approach consistently outperforms the most advanced methods on popular seasonal change detection datasets.
arXiv Detail & Related papers (2022-08-12T14:24:09Z) - Weakly Aligned Feature Fusion for Multimodal Object Detection [52.15436349488198]
multimodal data often suffer from the position shift problem, i.e., the image pair is not strictly aligned.
This problem makes it difficult to fuse multimodal features and puzzles the convolutional neural network (CNN) training.
In this article, we propose a general multimodal detector named aligned region CNN (AR-CNN) to tackle the position shift problem.
arXiv Detail & Related papers (2022-04-21T02:35:23Z) - AdaZoom: Adaptive Zoom Network for Multi-Scale Object Detection in Large
Scenes [57.969186815591186]
Detection in large-scale scenes is a challenging problem due to small objects and extreme scale variation.
We propose a novel Adaptive Zoom (AdaZoom) network as a selective magnifier with flexible shape and focal length to adaptively zoom the focus regions for object detection.
arXiv Detail & Related papers (2021-06-19T03:30:22Z) - Semantic Change Detection with Asymmetric Siamese Networks [71.28665116793138]
Given two aerial images, semantic change detection aims to locate the land-cover variations and identify their change types with pixel-wise boundaries.
This problem is vital in many earth vision related tasks, such as precise urban planning and natural resource management.
We present an asymmetric siamese network (ASN) to locate and identify semantic changes through feature pairs obtained from modules of widely different structures.
arXiv Detail & Related papers (2020-10-12T13:26:30Z) - Unsupervised Bidirectional Cross-Modality Adaptation via Deeply
Synergistic Image and Feature Alignment for Medical Image Segmentation [73.84166499988443]
We present a novel unsupervised domain adaptation framework, named as Synergistic Image and Feature Alignment (SIFA)
Our proposed SIFA conducts synergistic alignment of domains from both image and feature perspectives.
Experimental results on two different tasks demonstrate that our SIFA method is effective in improving segmentation performance on unlabeled target images.
arXiv Detail & Related papers (2020-02-06T13:49:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.