Applying VertexShuffle Toward 360-Degree Video Super-Resolution on
Focused-Icosahedral-Mesh
- URL: http://arxiv.org/abs/2106.11253v1
- Date: Mon, 21 Jun 2021 16:53:57 GMT
- Title: Applying VertexShuffle Toward 360-Degree Video Super-Resolution on
Focused-Icosahedral-Mesh
- Authors: Na Li and Yao Liu
- Abstract summary: We exploit Focused Icosahedral Mesh to represent a small area and construct matrices to rotate spherical content to the focused mesh area.
We also proposed a novel VertexShuffle operation that can significantly improve both the performance and the efficiency.
Our proposed spherical super-resolution model achieves significant benefits in terms of both performance and inference time.
- Score: 10.29596292902288
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the emerging of 360-degree image/video, augmented reality (AR) and
virtual reality (VR), the demand for analysing and processing spherical signals
get tremendous increase. However, plenty of effort paid on planar signals that
projected from spherical signals, which leading to some problems, e.g. waste of
pixels, distortion. Recent advances in spherical CNN have opened up the
possibility of directly analysing spherical signals. However, they pay
attention to the full mesh which makes it infeasible to deal with situations in
real-world application due to the extremely large bandwidth requirement. To
address the bandwidth waste problem associated with 360-degree video streaming
and save computation, we exploit Focused Icosahedral Mesh to represent a small
area and construct matrices to rotate spherical content to the focused mesh
area. We also proposed a novel VertexShuffle operation that can significantly
improve both the performance and the efficiency compared to the original
MeshConv Transpose operation introduced in UGSCNN. We further apply our
proposed methods on super resolution model, which is the first to propose a
spherical super-resolution model that directly operates on a mesh
representation of spherical pixels of 360-degree data. To evaluate our model,
we also collect a set of high-resolution 360-degree videos to generate a
spherical image dataset. Our experiments indicate that our proposed spherical
super-resolution model achieves significant benefits in terms of both
performance and inference time compared to the baseline spherical
super-resolution model that uses the simple MeshConv Transpose operation. In
summary, our model achieves great super-resolution performance on 360-degree
inputs, achieving 32.79 dB PSNR on average when super-resoluting 16x vertices
on the mesh.
Related papers
- Mesh2NeRF: Direct Mesh Supervision for Neural Radiance Field Representation and Generation [51.346733271166926]
Mesh2NeRF is an approach to derive ground-truth radiance fields from textured meshes for 3D generation tasks.
We validate the effectiveness of Mesh2NeRF across various tasks.
arXiv Detail & Related papers (2024-03-28T11:22:53Z) - HybridNeRF: Efficient Neural Rendering via Adaptive Volumetric Surfaces [71.1071688018433]
Neural radiance fields provide state-of-the-art view synthesis quality but tend to be slow to render.
We propose a method, HybridNeRF, that leverages the strengths of both representations by rendering most objects as surfaces.
We improve error rates by 15-30% while achieving real-time framerates (at least 36 FPS) for virtual-reality resolutions (2Kx2K)
arXiv Detail & Related papers (2023-12-05T22:04:49Z) - LightGaussian: Unbounded 3D Gaussian Compression with 15x Reduction and 200+ FPS [55.85673901231235]
We introduce LightGaussian, a method for transforming 3D Gaussians into a more compact format.
Inspired by Network Pruning, LightGaussian identifies Gaussians with minimal global significance on scene reconstruction.
LightGaussian achieves an average 15x compression rate while boosting FPS from 144 to 237 within the 3D-GS framework.
arXiv Detail & Related papers (2023-11-28T21:39:20Z) - OPDN: Omnidirectional Position-aware Deformable Network for
Omnidirectional Image Super-Resolution [18.138867445188293]
We propose a two-stage framework for 360deg omnidirectional image superresolution.
Our proposed method achieves superior performance and wins the NTIRE 2023 challenge of 360deg omnidirectional image super-resolution.
arXiv Detail & Related papers (2023-04-26T11:47:40Z) - Balanced Spherical Grid for Egocentric View Synthesis [6.518792457424123]
We present EgoNeRF, a practical solution to reconstruct large-scale real-world environments for VR assets.
Given a few seconds of casually captured 360 video, EgoNeRF can efficiently build neural radiance fields.
arXiv Detail & Related papers (2023-03-22T09:17:01Z) - NerfDiff: Single-image View Synthesis with NeRF-guided Distillation from
3D-aware Diffusion [107.67277084886929]
Novel view synthesis from a single image requires inferring occluded regions of objects and scenes whilst simultaneously maintaining semantic and physical consistency with the input.
We propose NerfDiff, which addresses this issue by distilling the knowledge of a 3D-aware conditional diffusion model (CDM) into NeRF through synthesizing and refining a set of virtual views at test time.
We further propose a novel NeRF-guided distillation algorithm that simultaneously generates 3D consistent virtual views from the CDM samples, and finetunes the NeRF based on the improved virtual views.
arXiv Detail & Related papers (2023-02-20T17:12:00Z) - Flow-based GAN for 3D Point Cloud Generation from a Single Image [16.04710129379503]
We introduce a hybrid explicit-implicit generative modeling scheme, which inherits the flow-based explicit generative models for sampling point clouds with arbitrary resolutions.
We evaluate on the large-scale synthetic dataset ShapeNet, with the experimental results demonstrating the superior performance of the proposed method.
arXiv Detail & Related papers (2022-10-08T17:58:20Z) - SphereSR: 360{\deg} Image Super-Resolution with Arbitrary Projection via
Continuous Spherical Image Representation [27.10716804733828]
We propose a novel framework to generate a continuous spherical image representation from an LR 360degimage.
Specifically, we first propose a feature extraction module that represents the spherical data based on icosahedron.
We then propose a spherical local implicit image function (SLIIF) to predict RGB values at the spherical coordinates.
arXiv Detail & Related papers (2021-12-13T10:16:51Z) - Robust 360-8PA: Redesigning The Normalized 8-point Algorithm for 360-FoV
Images [53.11097060367591]
We present a novel strategy for estimating an essential matrix from 360-FoV images in spherical projection.
We show that our normalization can increase the camera pose accuracy by about 20% without significantly overhead the time.
arXiv Detail & Related papers (2021-04-22T07:23:11Z) - Concentric Spherical GNN for 3D Representation Learning [53.45704095146161]
We propose a novel multi-resolution convolutional architecture for learning over concentric spherical feature maps.
Our hierarchical architecture is based on alternatively learning to incorporate both intra-sphere and inter-sphere information.
We demonstrate the effectiveness of our approach in improving state-of-the-art performance on 3D classification tasks with rotated data.
arXiv Detail & Related papers (2021-03-18T19:05:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.