MINER: Multiscale Implicit Neural Representations
- URL: http://arxiv.org/abs/2202.03532v1
- Date: Mon, 7 Feb 2022 21:49:33 GMT
- Title: MINER: Multiscale Implicit Neural Representations
- Authors: Vishwanath Saragadam, Jasper Tan, Guha Balakrishnan, Richard G.
Baraniuk, Ashok Veeraraghavan
- Abstract summary: We introduce a new neural signal representation designed for the efficient high-resolution representation of large-scale signals.
The key innovation in our multiscale implicit neural representation (MINER) is an internal representation via a Laplacian pyramid.
We demonstrate that it requires fewer than 25% of the parameters, 33% of the memory footprint, and 10% of the time of competing techniques such as ACORN to reach the same representation error.
- Score: 43.36327238440042
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce a new neural signal representation designed for the efficient
high-resolution representation of large-scale signals. The key innovation in
our multiscale implicit neural representation (MINER) is an internal
representation via a Laplacian pyramid, which provides a sparse multiscale
representation of the signal that captures orthogonal parts of the signal
across scales. We leverage the advantages of the Laplacian pyramid by
representing small disjoint patches of the pyramid at each scale with a tiny
MLP. This enables the capacity of the network to adaptively increase from
coarse to fine scales, and only represent parts of the signal with strong
signal energy. The parameters of each MLP are optimized from coarse-to-fine
scale which results in faster approximations at coarser scales, thereby
ultimately an extremely fast training process. We apply MINER to a range of
large-scale signal representation tasks, including gigapixel images and very
large point clouds, and demonstrate that it requires fewer than 25% of the
parameters, 33% of the memory footprint, and 10% of the computation time of
competing techniques such as ACORN to reach the same representation error.
Related papers
- Attention Beats Linear for Fast Implicit Neural Representation Generation [13.203243059083533]
We propose Attention-based Localized INR (ANR) composed of a localized attention layer (LAL) and a global representation vector.
With instance-specific representation and instance-agnostic ANR parameters, the target signals are well reconstructed as a continuous function.
arXiv Detail & Related papers (2024-07-22T03:52:18Z) - Versatile Neural Processes for Learning Implicit Neural Representations [57.090658265140384]
We propose Versatile Neural Processes (VNP), which largely increases the capability of approximating functions.
Specifically, we introduce a bottleneck encoder that produces fewer and informative context tokens, relieving the high computational cost.
We demonstrate the effectiveness of the proposed VNP on a variety of tasks involving 1D, 2D and 3D signals.
arXiv Detail & Related papers (2023-01-21T04:08:46Z) - Variable Bitrate Neural Fields [75.24672452527795]
We present a dictionary method for compressing feature grids, reducing their memory consumption by up to 100x.
We formulate the dictionary optimization as a vector-quantized auto-decoder problem which lets us learn end-to-end discrete neural representations in a space where no direct supervision is available.
arXiv Detail & Related papers (2022-06-15T17:58:34Z) - CoordX: Accelerating Implicit Neural Representation with a Split MLP
Architecture [2.6912336656165805]
Implicit neural representations with multi-layer perceptrons (MLPs) have recently gained prominence for a wide variety of tasks.
We propose a new split architecture, CoordX, to accelerate inference and training of coordinate-based representations.
We demonstrate a speedup of up to 2.92x compared to the baseline model for image, video, and 3D shape representation and rendering tasks.
arXiv Detail & Related papers (2022-01-28T21:30:42Z) - Meta-Learning Sparse Implicit Neural Representations [69.15490627853629]
Implicit neural representations are a promising new avenue of representing general signals.
Current approach is difficult to scale for a large number of signals or a data set.
We show that meta-learned sparse neural representations achieve a much smaller loss than dense meta-learned models.
arXiv Detail & Related papers (2021-10-27T18:02:53Z) - ACORN: Adaptive Coordinate Networks for Neural Scene Representation [40.04760307540698]
Current neural representations fail to accurately represent images at resolutions greater than a megapixel or 3D scenes with more than a few hundred thousand polygons.
We introduce a new hybrid implicit-explicit network architecture and training strategy that adaptively allocates resources during training and inference.
We demonstrate the first experiments that fit gigapixel images to nearly 40 dB peak signal-to-noise ratio.
arXiv Detail & Related papers (2021-05-06T16:21:38Z) - Modulated Periodic Activations for Generalizable Local Functional
Representations [113.64179351957888]
We present a new representation that generalizes to multiple instances and achieves state-of-the-art fidelity.
Our approach produces general functional representations of images, videos and shapes, and achieves higher reconstruction quality than prior works that are optimized for a single signal.
arXiv Detail & Related papers (2021-04-08T17:59:04Z) - Learned Initializations for Optimizing Coordinate-Based Neural
Representations [47.408295381897815]
Coordinate-based neural representations have shown significant promise as an alternative to discrete, array-based representations.
We propose applying standard meta-learning algorithms to learn the initial weight parameters for these fully-connected networks.
We explore these benefits across a variety of tasks, including representing 2D images, reconstructing CT scans, and recovering 3D shapes and scenes from 2D image observations.
arXiv Detail & Related papers (2020-12-03T18:59:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.