Parallel Discrete Convolutions on Adaptive Particle Representations of
Images
- URL: http://arxiv.org/abs/2112.03592v1
- Date: Tue, 7 Dec 2021 09:40:05 GMT
- Title: Parallel Discrete Convolutions on Adaptive Particle Representations of
Images
- Authors: Joel Jonsson, Bevan L. Cheeseman, Suryanarayana Maddu, Krzysztof
Gonciarz, Ivo F. Sbalzarini
- Abstract summary: We present data structures and algorithms for native implementations of discrete convolution operators over Adaptive Particle Representations.
The APR is a content-adaptive image representation that locally adapts the sampling resolution to the image signal.
We show that APR convolution naturally leads to scale-adaptive algorithms that efficiently parallelize on multi-core CPU and GPU architectures.
- Score: 2.362412515574206
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We present data structures and algorithms for native implementations of
discrete convolution operators over Adaptive Particle Representations (APR) of
images on parallel computer architectures. The APR is a content-adaptive image
representation that locally adapts the sampling resolution to the image signal.
It has been developed as an alternative to pixel representations for large,
sparse images as they typically occur in fluorescence microscopy. It has been
shown to reduce the memory and runtime costs of storing, visualizing, and
processing such images. This, however, requires that image processing natively
operates on APRs, without intermediately reverting to pixels. Designing
efficient and scalable APR-native image processing primitives, however, is
complicated by the APR's irregular memory structure. Here, we provide the
algorithmic building blocks required to efficiently and natively process APR
images using a wide range of algorithms that can be formulated in terms of
discrete convolutions. We show that APR convolution naturally leads to
scale-adaptive algorithms that efficiently parallelize on multi-core CPU and
GPU architectures. We quantify the speedups in comparison to pixel-based
algorithms and convolutions on evenly sampled data. We achieve pixel-equivalent
throughputs of up to 1 TB/s on a single Nvidia GeForce RTX 2080 gaming GPU,
requiring up to two orders of magnitude less memory than a pixel-based
implementation.
Related papers
- Accelerating Image Super-Resolution Networks with Pixel-Level Classification [29.010136088811137]
Pixel-level for Single Image SuperResolution is a novel method designed to distribute computational resources adaptively at the pixel level.
Our method allows for performance and computational cost balance during inference without re-training.
arXiv Detail & Related papers (2024-07-31T08:53:10Z) - Image-GS: Content-Adaptive Image Representation via 2D Gaussians [55.15950594752051]
We propose Image-GS, a content-adaptive image representation.
Using anisotropic 2D Gaussians as the basis, Image-GS shows high memory efficiency, supports fast random access, and offers a natural level of detail stack.
General efficiency and fidelity of Image-GS are validated against several recent neural image representations and industry-standard texture compressors.
We hope this research offers insights for developing new applications that require adaptive quality and resource control, such as machine perception, asset streaming, and content generation.
arXiv Detail & Related papers (2024-07-02T00:45:21Z) - Parameter-Inverted Image Pyramid Networks [49.35689698870247]
We propose a novel network architecture known as the Inverted Image Pyramid Networks (PIIP)
Our core idea is to use models with different parameter sizes to process different resolution levels of the image pyramid.
PIIP achieves superior performance in tasks such as object detection, segmentation, and image classification.
arXiv Detail & Related papers (2024-06-06T17:59:10Z) - Mapping Image Transformations Onto Pixel Processor Arrays [4.857223862405921]
Pixel Processor Arrays (PPA) present a new vision sensor/processor architecture consisting of a SIMD array of processor elements.
We demonstrate how various image transformations, including shearing, rotation and scaling, can be performed directly upon a PPA.
arXiv Detail & Related papers (2024-03-25T17:56:41Z) - ImageSig: A signature transform for ultra-lightweight image recognition [0.0]
ImageSig is based on computing signatures and does not require a convolutional structure or an attention-based encoder.
ImageSig shows unprecedented performance on hardware such as Raspberry Pi and Jetson-nano.
arXiv Detail & Related papers (2022-05-13T23:48:32Z) - Fast and High-Quality Image Denoising via Malleable Convolutions [72.18723834537494]
We present Malleable Convolution (MalleConv), as an efficient variant of dynamic convolution.
Unlike previous works, MalleConv generates a much smaller set of spatially-varying kernels from input.
We also build an efficient denoising network using MalleConv, coined as MalleNet.
arXiv Detail & Related papers (2022-01-02T18:35:20Z) - Spatially-Adaptive Pixelwise Networks for Fast Image Translation [57.359250882770525]
We introduce a new generator architecture, aimed at fast and efficient high-resolution image-to-image translation.
We use pixel-wise networks; that is, each pixel is processed independently of others.
Our model is up to 18x faster than state-of-the-art baselines.
arXiv Detail & Related papers (2020-12-05T10:02:03Z) - Adversarial Generation of Continuous Images [31.92891885615843]
In this paper, we propose two novel architectural techniques for building INR-based image decoders.
We use them to build a state-of-the-art continuous image GAN.
Our proposed INR-GAN architecture improves the performance of continuous image generators by several times.
arXiv Detail & Related papers (2020-11-24T11:06:40Z) - Locally Masked Convolution for Autoregressive Models [107.4635841204146]
LMConv is a simple modification to the standard 2D convolution that allows arbitrary masks to be applied to the weights at each location in the image.
We learn an ensemble of distribution estimators that share parameters but differ in generation order, achieving improved performance on whole-image density estimation.
arXiv Detail & Related papers (2020-06-22T17:59:07Z) - Visual Transformers: Token-based Image Representation and Processing for
Computer Vision [67.55770209540306]
Visual Transformer ( VT) operates in a semantic token space, judiciously attending to different image parts based on context.
Using an advanced training recipe, our VTs significantly outperform their convolutional counterparts.
For semantic segmentation on LIP and COCO-stuff, VT-based feature pyramid networks (FPN) achieve 0.35 points higher mIoU while reducing the FPN module's FLOPs by 6.5x.
arXiv Detail & Related papers (2020-06-05T20:49:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.