NeX: Real-time View Synthesis with Neural Basis Expansion
- URL: http://arxiv.org/abs/2103.05606v1
- Date: Tue, 9 Mar 2021 18:27:27 GMT
- Title: NeX: Real-time View Synthesis with Neural Basis Expansion
- Authors: Suttisak Wizadwongsa, Pakkapon Phongthawee, Jiraphon Yenphraphai,
Supasorn Suwajanakorn
- Abstract summary: We present NeX, a new approach to novel view synthesis based on enhancements of multiplane image (MPI)
Our technique models view-dependent effects by instead parameterizing each pixel as a linear combination of basis functions learned from a neural network.
- Score: 1.471992435706872
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present NeX, a new approach to novel view synthesis based on enhancements
of multiplane image (MPI) that can reproduce next-level view-dependent effects
-- in real time. Unlike traditional MPI that uses a set of simple RGB$\alpha$
planes, our technique models view-dependent effects by instead parameterizing
each pixel as a linear combination of basis functions learned from a neural
network. Moreover, we propose a hybrid implicit-explicit modeling strategy that
improves upon fine detail and produces state-of-the-art results. Our method is
evaluated on benchmark forward-facing datasets as well as our newly-introduced
dataset designed to test the limit of view-dependent modeling with
significantly more challenging effects such as rainbow reflections on a CD. Our
method achieves the best overall scores across all major metrics on these
datasets with more than 1000$\times$ faster rendering time than the state of
the art. For real-time demos, visit https://nex-mpi.github.io/
Related papers
- HybridNeRF: Efficient Neural Rendering via Adaptive Volumetric Surfaces [71.1071688018433]
Neural radiance fields provide state-of-the-art view synthesis quality but tend to be slow to render.
We propose a method, HybridNeRF, that leverages the strengths of both representations by rendering most objects as surfaces.
We improve error rates by 15-30% while achieving real-time framerates (at least 36 FPS) for virtual-reality resolutions (2Kx2K)
arXiv Detail & Related papers (2023-12-05T22:04:49Z) - DNS SLAM: Dense Neural Semantic-Informed SLAM [92.39687553022605]
DNS SLAM is a novel neural RGB-D semantic SLAM approach featuring a hybrid representation.
Our method integrates multi-view geometry constraints with image-based feature extraction to improve appearance details.
Our experimental results achieve state-of-the-art performance on both synthetic data and real-world data tracking.
arXiv Detail & Related papers (2023-11-30T21:34:44Z) - Fast Non-Rigid Radiance Fields from Monocularized Data [66.74229489512683]
This paper proposes a new method for full 360deg inward-facing novel view synthesis of non-rigidly deforming scenes.
At the core of our method are 1) An efficient deformation module that decouples the processing of spatial and temporal information for accelerated training and inference; and 2) A static module representing the canonical scene as a fast hash-encoded neural radiance field.
In both cases, our method is significantly faster than previous methods, converging in less than 7 minutes and achieving real-time framerates at 1K resolution, while obtaining a higher visual accuracy for generated novel views.
arXiv Detail & Related papers (2022-12-02T18:51:10Z) - Generalizable Patch-Based Neural Rendering [46.41746536545268]
We propose a new paradigm for learning models that can synthesize novel views of unseen scenes.
Our method is capable of predicting the color of a target ray in a novel scene directly, just from a collection of patches sampled from the scene.
We show that our approach outperforms the state-of-the-art on novel view synthesis of unseen scenes even when being trained with considerably less data than prior work.
arXiv Detail & Related papers (2022-07-21T17:57:04Z) - RTMV: A Ray-Traced Multi-View Synthetic Dataset for Novel View Synthesis [104.53930611219654]
We present a large-scale synthetic dataset for novel view synthesis consisting of 300k images rendered from nearly 2000 complex scenes.
The dataset is orders of magnitude larger than existing synthetic datasets for novel view synthesis.
Using 4 distinct sources of high-quality 3D meshes, the scenes of our dataset exhibit challenging variations in camera views, lighting, shape, materials, and textures.
arXiv Detail & Related papers (2022-05-14T13:15:32Z) - Learning Dynamic View Synthesis With Few RGBD Cameras [60.36357774688289]
We propose to utilize RGBD cameras to synthesize free-viewpoint videos of dynamic indoor scenes.
We generate point clouds from RGBD frames and then render them into free-viewpoint videos via a neural feature.
We introduce a simple Regional Depth-Inpainting module that adaptively inpaints missing depth values to render complete novel views.
arXiv Detail & Related papers (2022-04-22T03:17:35Z) - Towards Robustness of Neural Networks [0.0]
We introduce ImageNet-A/O and ImageNet-R as well as a synthetic environment and testing suite we called CAOS.
All of the datasets were created for testing robustness and measuring progress in robustness.
We build off of simple baselines in the form of Maximum Logit, and Typicality Score as well as create a novel data augmentation method in the form of DeepAugment.
arXiv Detail & Related papers (2021-12-30T19:41:10Z) - Spatial-Separated Curve Rendering Network for Efficient and
High-Resolution Image Harmonization [59.19214040221055]
We propose a novel spatial-separated curve rendering network (S$2$CRNet) for efficient and high-resolution image harmonization.
The proposed method reduces more than 90% parameters compared with previous methods.
Our method can work smoothly on higher resolution images in real-time which is more than 10$times$ faster than the existing methods.
arXiv Detail & Related papers (2021-09-13T07:20:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.