Quaternion-Based Robust PCA for Efficient Moving Target Detection and Background Recovery in Color Videos
- URL: http://arxiv.org/abs/2507.19730v1
- Date: Sat, 26 Jul 2025 01:05:03 GMT
- Title: Quaternion-Based Robust PCA for Efficient Moving Target Detection and Background Recovery in Color Videos
- Authors: Liyang Wang, Shiqian Wu, Shun Fang, Qile Zhu, Jiaxin Wu, Sos Again,
- Abstract summary: Quaternion-based RPCA (QRPCA) is a promising unsupervised paradigm for color image processing.<n>We propose the universal QRPCA framework, which achieves a balance in simultaneously segmenting targets and recovering backgrounds from color videos.<n>We also expand to uQRPCA+ by introducing the Color Rank-1 Batch (CR1B) method to further process and obtain the ideal low-rank background.
- Score: 6.948496559712165
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Moving target detection is a challenging computer vision task aimed at generating accurate segmentation maps in diverse in-the-wild color videos captured by static cameras. If backgrounds and targets can be simultaneously extracted and recombined, such synthetic data can significantly enrich annotated in-the-wild datasets and enhance the generalization ability of deep models. Quaternion-based RPCA (QRPCA) is a promising unsupervised paradigm for color image processing. However, in color video processing, Quaternion Singular Value Decomposition (QSVD) incurs high computational costs, and rank-1 quaternion matrix fails to yield rank-1 color channels. In this paper, we reduce the computational complexity of QSVD to o(1) by utilizing a quaternion Riemannian manifold. Furthermor, we propose the universal QRPCA (uQRPCA) framework, which achieves a balance in simultaneously segmenting targets and recovering backgrounds from color videos. Moreover, we expand to uQRPCA+ by introducing the Color Rank-1 Batch (CR1B) method to further process and obtain the ideal low-rank background across color channels. Experiments demonstrate our uQRPCA+ achieves State Of The Art (SOTA) performance on moving target detection and background recovery tasks compared to existing open-source methods. Our implementation is publicly available on GitHub at https://github.com/Ruchtech/uQRPCA
Related papers
- SIRST-5K: Exploring Massive Negatives Synthesis with Self-supervised
Learning for Robust Infrared Small Target Detection [53.19618419772467]
Single-frame infrared small target (SIRST) detection aims to recognize small targets from clutter backgrounds.
With the development of Transformer, the scale of SIRST models is constantly increasing.
With a rich diversity of infrared small target data, our algorithm significantly improves the model performance and convergence speed.
arXiv Detail & Related papers (2024-03-08T16:14:54Z) - RGB-based Category-level Object Pose Estimation via Decoupled Metric
Scale Recovery [72.13154206106259]
We propose a novel pipeline that decouples the 6D pose and size estimation to mitigate the influence of imperfect scales on rigid transformations.
Specifically, we leverage a pre-trained monocular estimator to extract local geometric information.
A separate branch is designed to directly recover the metric scale of the object based on category-level statistics.
arXiv Detail & Related papers (2023-09-19T02:20:26Z) - $L_{2,1}$-Norm Regularized Quaternion Matrix Completion Using Sparse
Representation and Quaternion QR Decomposition [7.344370881751104]
We propose a method based on quaternion Qatar Riyal decomposition (QQR) and quaternion $L_2,1$-norm called QLNM-QQR.
This new approach reduces computational complexity by avoiding the need to calculate the QSVD of large quaternion matrices.
We also present two improvements to the QLNM-QQR method: an enhanced version called IRQLNM-QQR that uses iteratively reweighted quaternion $L_2,1$-norm minimization and a method called QLNM-QQR-SR that
arXiv Detail & Related papers (2023-09-07T15:08:12Z) - A New Low-Rank Learning Robust Quaternion Tensor Completion Method for
Color Video Inpainting Problem and Fast Algorithms [0.0]
We present a new robust quaternion tensor completion (RQTC) model to solve this challenging problem and derive the exact recovery theory.
In numerical experiments, the proposed methods successfully recover color videos with eliminating color contamination and keeping the continuity of video scenery.
arXiv Detail & Related papers (2023-06-16T07:15:38Z) - Dual Swin-Transformer based Mutual Interactive Network for RGB-D Salient
Object Detection [67.33924278729903]
In this work, we propose Dual Swin-Transformer based Mutual Interactive Network.
We adopt Swin-Transformer as the feature extractor for both RGB and depth modality to model the long-range dependencies in visual inputs.
Comprehensive experiments on five standard RGB-D SOD benchmark datasets demonstrate the superiority of the proposed DTMINet method.
arXiv Detail & Related papers (2022-06-07T08:35:41Z) - Multi-Source Fusion and Automatic Predictor Selection for Zero-Shot
Video Object Segmentation [86.94578023985677]
We propose a novel multi-source fusion network for zero-shot video object segmentation.
The proposed model achieves compelling performance against the state-of-the-arts.
arXiv Detail & Related papers (2021-08-11T07:37:44Z) - Quaternion higher-order singular value decomposition and its
applications in color image processing [1.1929584800629671]
We generalize the HOSVD to the quaternion domain and define quaternion-based HOSVD (QHOSVD)
Due to the non-commutability of quaternion multiplication, QHOSVD is not a trivial extension of the HOSVD.
We present two applications of the defined QHOSVD in color image processing: multi_focus color image fusion and color image denoising.
arXiv Detail & Related papers (2021-01-02T03:54:56Z) - Generalized Two-Dimensional Quaternion Principal Component Analysis with
Weighting for Color Image Recognition [1.516937009186805]
One of the most powerful methods of color image recognition is the two-dimensional principle component analysis (2DQPCA) approach.
In this paper, a generalized 2DQPCA approach with weighting is presented with imposing $L_p$ norms on both constraint and objective functions.
The numerical results based on the real face databases validate that the newly proposed method is robust to noise and performs better than the state-of-the-art 2DQPCA-based algorithms and four prominent deep learning methods.
arXiv Detail & Related papers (2020-10-04T03:37:23Z) - Uncertainty Inspired RGB-D Saliency Detection [70.50583438784571]
We propose the first framework to employ uncertainty for RGB-D saliency detection by learning from the data labeling process.
Inspired by the saliency data labeling process, we propose a generative architecture to achieve probabilistic RGB-D saliency detection.
Results on six challenging RGB-D benchmark datasets show our approach's superior performance in learning the distribution of saliency maps.
arXiv Detail & Related papers (2020-09-07T13:01:45Z) - Progressively Guided Alternate Refinement Network for RGB-D Salient
Object Detection [63.18846475183332]
We aim to develop an efficient and compact deep network for RGB-D salient object detection.
We propose a progressively guided alternate refinement network to refine it.
Our model outperforms existing state-of-the-art approaches by a large margin.
arXiv Detail & Related papers (2020-08-17T02:55:06Z) - Full Quaternion Representation of Color images: A Case Study on
QSVD-based Color Image Compression [0.38073142980732994]
We propose an approach for representing color images with full quaternion numbers.
An autoencoder neural network is used to generate a global model for transforming a color image into a full quaternion matrix.
arXiv Detail & Related papers (2020-07-19T19:13:21Z) - Bi-directional Cross-Modality Feature Propagation with
Separation-and-Aggregation Gate for RGB-D Semantic Segmentation [59.94819184452694]
Depth information has proven to be a useful cue in the semantic segmentation of RGBD images for providing a geometric counterpart to the RGB representation.
Most existing works simply assume that depth measurements are accurate and well-aligned with the RGB pixels and models the problem as a cross-modal feature fusion.
In this paper, we propose a unified and efficient Crossmodality Guided to not only effectively recalibrate RGB feature responses, but also to distill accurate depth information via multiple stages and aggregate the two recalibrated representations alternatively.
arXiv Detail & Related papers (2020-07-17T18:35:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.