MM-PCQA: Multi-Modal Learning for No-reference Point Cloud Quality
Assessment
- URL: http://arxiv.org/abs/2209.00244v2
- Date: Mon, 24 Apr 2023 08:46:31 GMT
- Title: MM-PCQA: Multi-Modal Learning for No-reference Point Cloud Quality
Assessment
- Authors: Zicheng Zhang, Wei Sun, Xiongkuo Min, Quan Zhou, Jun He, Qiyuan Wang,
and Guangtao Zhai
- Abstract summary: We propose a novel no-reference point cloud quality assessment (NR-PCQA) metric in a multi-modal fashion.
In specific, we split the point clouds into sub-models to represent local geometry distortions such as point shift and down-sampling.
To achieve the goals, the sub-models and projected images are encoded with point-based and image-based neural networks.
- Score: 32.495387943305204
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The visual quality of point clouds has been greatly emphasized since the
ever-increasing 3D vision applications are expected to provide cost-effective
and high-quality experiences for users. Looking back on the development of
point cloud quality assessment (PCQA) methods, the visual quality is usually
evaluated by utilizing single-modal information, i.e., either extracted from
the 2D projections or 3D point cloud. The 2D projections contain rich texture
and semantic information but are highly dependent on viewpoints, while the 3D
point clouds are more sensitive to geometry distortions and invariant to
viewpoints. Therefore, to leverage the advantages of both point cloud and
projected image modalities, we propose a novel no-reference point cloud quality
assessment (NR-PCQA) metric in a multi-modal fashion. In specific, we split the
point clouds into sub-models to represent local geometry distortions such as
point shift and down-sampling. Then we render the point clouds into 2D image
projections for texture feature extraction. To achieve the goals, the
sub-models and projected images are encoded with point-based and image-based
neural networks. Finally, symmetric cross-modal attention is employed to fuse
multi-modal quality-aware information. Experimental results show that our
approach outperforms all compared state-of-the-art methods and is far ahead of
previous NR-PCQA methods, which highlights the effectiveness of the proposed
method. The code is available at https://github.com/zzc-1998/MM-PCQA.
Related papers
- No-Reference Point Cloud Quality Assessment via Graph Convolutional Network [89.12589881881082]
Three-dimensional (3D) point cloud, as an emerging visual media format, is increasingly favored by consumers.
Point clouds inevitably suffer from quality degradation and information loss through multimedia communication systems.
We propose a novel no-reference PCQA method by using a graph convolutional network (GCN) to characterize the mutual dependencies of multi-view 2D projected image contents.
arXiv Detail & Related papers (2024-11-12T11:39:05Z) - Contrastive Pre-Training with Multi-View Fusion for No-Reference Point Cloud Quality Assessment [49.36799270585947]
No-reference point cloud quality assessment (NR-PCQA) aims to automatically evaluate the perceptual quality of distorted point clouds without available reference.
We propose a novel contrastive pre-training framework tailored for PCQA (CoPA)
Our method outperforms the state-of-the-art PCQA methods on popular benchmarks.
arXiv Detail & Related papers (2024-03-15T07:16:07Z) - Activating Frequency and ViT for 3D Point Cloud Quality Assessment
without Reference [0.49157446832511503]
We propose no-reference quality metric of a given 3D-PC.
To map the input attributes to quality score, we use a light-weight hybrid deep model; combined of Deformable Convolutional Network (DCN) and Vision Transformers (ViT)
The results show that our approach outperforms state-of-the-art NR-PCQA measures and even some FR-PCQA on PointXR.
arXiv Detail & Related papers (2023-12-10T19:13:34Z) - Zero-Shot Point Cloud Registration [94.39796531154303]
ZeroReg is the first zero-shot point cloud registration approach that eliminates the need for training on point cloud datasets.
The cornerstone of ZeroReg is the novel transfer of image features from keypoints to the point cloud, enriched by aggregating information from 3D geometric neighborhoods.
On benchmarks such as 3DMatch, 3DLoMatch, and ScanNet, ZeroReg achieves impressive Recall Ratios (RR) of over 84%, 46%, and 75%, respectively.
arXiv Detail & Related papers (2023-12-05T11:33:16Z) - Simple Baselines for Projection-based Full-reference and No-reference
Point Cloud Quality Assessment [60.2709006613171]
We propose simple baselines for projection-based point cloud quality assessment (PCQA)
We use multi-projections obtained via a common cube-like projection process from the point clouds for both full-reference (FR) and no-reference (NR) PCQA tasks.
Taking part in the ICIP 2023 PCVQA Challenge, we succeeded in achieving the top spot in four out of the five competition tracks.
arXiv Detail & Related papers (2023-10-26T04:42:57Z) - Blind Quality Assessment of 3D Dense Point Clouds with Structure Guided
Resampling [71.68672977990403]
We propose an objective point cloud quality index with Structure Guided Resampling (SGR) to automatically evaluate the perceptually visual quality of 3D dense point clouds.
The proposed SGR is a general-purpose blind quality assessment method without the assistance of any reference information.
arXiv Detail & Related papers (2022-08-31T02:42:55Z) - Evaluating Point Cloud from Moving Camera Videos: A No-Reference Metric [58.309735075960745]
This paper explores the way of dealing with point cloud quality assessment (PCQA) tasks via video quality assessment (VQA) methods.
We generate the captured videos by rotating the camera around the point clouds through several circular pathways.
We extract both spatial and temporal quality-aware features from the selected key frames and the video clips through using trainable 2D-CNN and pre-trained 3D-CNN models.
arXiv Detail & Related papers (2022-08-30T08:59:41Z) - A No-reference Quality Assessment Metric for Point Cloud Based on
Captured Video Sequences [40.46566408312466]
We propose a no-reference quality assessment metric for colored point cloud based on captured video sequences.
The experimental results show that our method outperforms most of the state-of-the-art full-reference and no-reference PCQA metrics.
arXiv Detail & Related papers (2022-06-09T06:42:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.