Applying VertexShuffle Toward 360-Degree Video Super-Resolution on
Focused-Icosahedral-Mesh
- URL: http://arxiv.org/abs/2106.11253v1
- Date: Mon, 21 Jun 2021 16:53:57 GMT
- Title: Applying VertexShuffle Toward 360-Degree Video Super-Resolution on
Focused-Icosahedral-Mesh
- Authors: Na Li and Yao Liu
- Abstract summary: We exploit Focused Icosahedral Mesh to represent a small area and construct matrices to rotate spherical content to the focused mesh area.
We also proposed a novel VertexShuffle operation that can significantly improve both the performance and the efficiency.
Our proposed spherical super-resolution model achieves significant benefits in terms of both performance and inference time.
- Score: 10.29596292902288
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the emerging of 360-degree image/video, augmented reality (AR) and
virtual reality (VR), the demand for analysing and processing spherical signals
get tremendous increase. However, plenty of effort paid on planar signals that
projected from spherical signals, which leading to some problems, e.g. waste of
pixels, distortion. Recent advances in spherical CNN have opened up the
possibility of directly analysing spherical signals. However, they pay
attention to the full mesh which makes it infeasible to deal with situations in
real-world application due to the extremely large bandwidth requirement. To
address the bandwidth waste problem associated with 360-degree video streaming
and save computation, we exploit Focused Icosahedral Mesh to represent a small
area and construct matrices to rotate spherical content to the focused mesh
area. We also proposed a novel VertexShuffle operation that can significantly
improve both the performance and the efficiency compared to the original
MeshConv Transpose operation introduced in UGSCNN. We further apply our
proposed methods on super resolution model, which is the first to propose a
spherical super-resolution model that directly operates on a mesh
representation of spherical pixels of 360-degree data. To evaluate our model,
we also collect a set of high-resolution 360-degree videos to generate a
spherical image dataset. Our experiments indicate that our proposed spherical
super-resolution model achieves significant benefits in terms of both
performance and inference time compared to the baseline spherical
super-resolution model that uses the simple MeshConv Transpose operation. In
summary, our model achieves great super-resolution performance on 360-degree
inputs, achieving 32.79 dB PSNR on average when super-resoluting 16x vertices
on the mesh.
Related papers
- 3D Gaussian Splatting with Normal Information for Mesh Extraction and Improved Rendering [8.59572577251833]
We propose a novel regularization method using the gradients of a signed distance function estimated from the Gaussians.
We demonstrate the effectiveness of our approach on datasets such as Mip-NeRF360, Tanks and Temples, and Deep-Blending.
arXiv Detail & Related papers (2025-01-14T18:40:33Z) - Revisiting 360 Depth Estimation with PanoGabor: A New Fusion Perspective [89.53522682640337]
We propose an oriented distortion-aware Gabor Fusion framework (PGFuse) to address the above challenges.
To address the reintroduced distortions, we design a linear latitude-aware distortion representation method to generate customized, distortion-aware Gabor filters.
Considering the orientation sensitivity of the Gabor transform, we introduce a spherical gradient constraint to stabilize this sensitivity.
arXiv Detail & Related papers (2024-08-29T02:58:35Z) - Mesh2NeRF: Direct Mesh Supervision for Neural Radiance Field Representation and Generation [51.346733271166926]
Mesh2NeRF is an approach to derive ground-truth radiance fields from textured meshes for 3D generation tasks.
We validate the effectiveness of Mesh2NeRF across various tasks.
arXiv Detail & Related papers (2024-03-28T11:22:53Z) - Hybrid Explicit Representation for Ultra-Realistic Head Avatars [55.829497543262214]
We introduce a novel approach to creating ultra-realistic head avatars and rendering them in real-time.
UV-mapped 3D mesh is utilized to capture sharp and rich textures on smooth surfaces, while 3D Gaussian Splatting is employed to represent complex geometric structures.
Experiments that our modeled results exceed those of state-of-the-art approaches.
arXiv Detail & Related papers (2024-03-18T04:01:26Z) - HybridNeRF: Efficient Neural Rendering via Adaptive Volumetric Surfaces [71.1071688018433]
Neural radiance fields provide state-of-the-art view synthesis quality but tend to be slow to render.
We propose a method, HybridNeRF, that leverages the strengths of both representations by rendering most objects as surfaces.
We improve error rates by 15-30% while achieving real-time framerates (at least 36 FPS) for virtual-reality resolutions (2Kx2K)
arXiv Detail & Related papers (2023-12-05T22:04:49Z) - LightGaussian: Unbounded 3D Gaussian Compression with 15x Reduction and 200+ FPS [55.85673901231235]
We introduce LightGaussian, a method for transforming 3D Gaussians into a more compact format.
Inspired by Network Pruning, LightGaussian identifies Gaussians with minimal global significance on scene reconstruction.
LightGaussian achieves an average 15x compression rate while boosting FPS from 144 to 237 within the 3D-GS framework.
arXiv Detail & Related papers (2023-11-28T21:39:20Z) - OPDN: Omnidirectional Position-aware Deformable Network for
Omnidirectional Image Super-Resolution [18.138867445188293]
We propose a two-stage framework for 360deg omnidirectional image superresolution.
Our proposed method achieves superior performance and wins the NTIRE 2023 challenge of 360deg omnidirectional image super-resolution.
arXiv Detail & Related papers (2023-04-26T11:47:40Z) - Balanced Spherical Grid for Egocentric View Synthesis [6.518792457424123]
We present EgoNeRF, a practical solution to reconstruct large-scale real-world environments for VR assets.
Given a few seconds of casually captured 360 video, EgoNeRF can efficiently build neural radiance fields.
arXiv Detail & Related papers (2023-03-22T09:17:01Z) - SphereSR: 360{\deg} Image Super-Resolution with Arbitrary Projection via
Continuous Spherical Image Representation [27.10716804733828]
We propose a novel framework to generate a continuous spherical image representation from an LR 360degimage.
Specifically, we first propose a feature extraction module that represents the spherical data based on icosahedron.
We then propose a spherical local implicit image function (SLIIF) to predict RGB values at the spherical coordinates.
arXiv Detail & Related papers (2021-12-13T10:16:51Z) - Robust 360-8PA: Redesigning The Normalized 8-point Algorithm for 360-FoV
Images [53.11097060367591]
We present a novel strategy for estimating an essential matrix from 360-FoV images in spherical projection.
We show that our normalization can increase the camera pose accuracy by about 20% without significantly overhead the time.
arXiv Detail & Related papers (2021-04-22T07:23:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.