Deep Contrastive Patch-Based Subspace Learning for Camera Image Signal
Processing
- URL: http://arxiv.org/abs/2104.00253v4
- Date: Tue, 3 Oct 2023 14:48:16 GMT
- Title: Deep Contrastive Patch-Based Subspace Learning for Camera Image Signal
Processing
- Authors: Yunhao Yang, Yi Wang, Chandrajit Bajaj
- Abstract summary: We present a specific patch-based, local subspace deep neural network that improves Camera ISP to be robust to heterogeneous artifacts.
We call our three-fold deep-trained model the Patch Subspace Learning Autoencoder (PSL-AE)
PSL-AE encodes patches extracted from noisy a nd clean image pairs, with different artifact types or distortion levels, by contrastive learning.
- Score: 5.678834480723395
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Camera Image Signal Processing (ISP) pipelines can get appealing results in
different image signal processing tasks. Nonetheless, the majority of these
methods, including those employing an encoder-decoder deep architecture for the
task, typically utilize a uniform filter applied consistently across the entire
image. However, it is natural to view a camera image as heterogeneous, as the
color intensity and the artificial noise are distributed vastly differently,
even across the two-dimensional domain of a single image. Varied Moire ringing,
motion blur, color-bleaching, or lens-based projection distortions can all
potentially lead to a heterogeneous image artifact filtering problem. In this
paper, we present a specific patch-based, local subspace deep neural network
that improves Camera ISP to be robust to heterogeneous artifacts (especially
image denoising). We call our three-fold deep-trained model the Patch Subspace
Learning Autoencoder (PSL-AE). The PSL-AE model does not make assumptions
regarding uniform levels of image distortion. Instead, it first encodes patches
extracted from noisy a nd clean image pairs, with different artifact types or
distortion levels, by contrastive learning. Then, the patches of each image are
encoded into corresponding soft clusters within their suitable latent
sub-space, utilizing a prior mixture model. Furthermore, the decoders undergo
training in an unsupervised manner, specifically trained for the image patches
present in each cluster. The experiments highlight the adaptability and
efficacy through enhanced heterogeneous filtering, both from synthesized
artifacts but also realistic SIDD image pairs.
Related papers
- Deep Learning Based Speckle Filtering for Polarimetric SAR Images. Application to Sentinel-1 [51.404644401997736]
We propose a complete framework to remove speckle in polarimetric SAR images using a convolutional neural network.
Experiments show that the proposed approach offers exceptional results in both speckle reduction and resolution preservation.
arXiv Detail & Related papers (2024-08-28T10:07:17Z) - Learning deep illumination-robust features from multispectral filter array images [0.5439020425819]
This paper presents an original approach to learn discriminant and illumination-robust features from directly from raw images.
Experiments on MS image classification show that our approach outperforms both handcrafted and recent deep learning-based methods.
arXiv Detail & Related papers (2024-07-22T08:35:41Z) - Neural Spline Fields for Burst Image Fusion and Layer Separation [40.9442467471977]
We propose a versatile intermediate representation: a two-layer alpha-composited image plus flow model constructed with neural spline fields.
Our method is able to jointly fuse a burst image capture into one high-resolution reconstruction and decompose it into transmission and obstruction layers.
We find that, with no post-processing steps or learned priors, our generalizable model is able to outperform existing dedicated single-image and multi-view obstruction removal approaches.
arXiv Detail & Related papers (2023-12-21T18:54:19Z) - Aperture Diffraction for Compact Snapshot Spectral Imaging [27.321750056840706]
We demonstrate a compact, cost-effective snapshot spectral imaging system named Aperture Diffraction Imaging Spectrometer (ADIS)
A new optical design that each point in the object space is multiplexed to discrete encoding locations on the mosaic filter sensor is introduced.
The Cascade Shift-Shuffle Spectral Transformer (CSST) with strong perception of the diffraction degeneration is designed to solve a sparsity-constrained inverse problem.
arXiv Detail & Related papers (2023-09-27T16:48:46Z) - Multiscale Representation for Real-Time Anti-Aliasing Neural Rendering [84.37776381343662]
Mip-NeRF proposes a multiscale representation as a conical frustum to encode scale information.
We propose mip voxel grids (Mip-VoG), an explicit multiscale representation for real-time anti-aliasing rendering.
Our approach is the first to offer multiscale training and real-time anti-aliasing rendering simultaneously.
arXiv Detail & Related papers (2023-04-20T04:05:22Z) - Parallax-Tolerant Unsupervised Deep Image Stitching [57.76737888499145]
We propose UDIS++, a parallax-tolerant unsupervised deep image stitching technique.
First, we propose a robust and flexible warp to model the image registration from global homography to local thin-plate spline motion.
To further eliminate the parallax artifacts, we propose to composite the stitched image seamlessly by unsupervised learning for seam-driven composition masks.
arXiv Detail & Related papers (2023-02-16T10:40:55Z) - Recursive Self-Improvement for Camera Image and Signal Processing
Pipeline [6.318974730864278]
Current camera image and signal processing pipelines (ISPs) tend to apply a single filter that is uniformly applied to the entire image.
This despite the fact that most acquired camera images have spatially heterogeneous artifacts.
We present a deep reinforcement learning model that works in learned latent subspaces.
arXiv Detail & Related papers (2021-11-15T02:23:40Z) - Ensembling with Deep Generative Views [72.70801582346344]
generative models can synthesize "views" of artificial images that mimic real-world variations, such as changes in color or pose.
Here, we investigate whether such views can be applied to real images to benefit downstream analysis tasks such as image classification.
We use StyleGAN2 as the source of generative augmentations and investigate this setup on classification tasks involving facial attributes, cat faces, and cars.
arXiv Detail & Related papers (2021-04-29T17:58:35Z) - Beyond Joint Demosaicking and Denoising: An Image Processing Pipeline
for a Pixel-bin Image Sensor [0.883717274344425]
Pixel binning is considered one of the most prominent solutions to tackle the hardware limitation of smartphone cameras.
In this paper, we tackle the challenges of joint demosaicing and denoising (JDD) on such an image sensor by introducing a novel learning-based method.
The proposed method is guided by a multi-term objective function, including two novel perceptual losses to produce visually plausible images.
arXiv Detail & Related papers (2021-04-19T15:41:28Z) - Designing a Practical Degradation Model for Deep Blind Image
Super-Resolution [134.9023380383406]
Single image super-resolution (SISR) methods would not perform well if the assumed degradation model deviates from those in real images.
This paper proposes to design a more complex but practical degradation model that consists of randomly shuffled blur, downsampling and noise degradations.
arXiv Detail & Related papers (2021-03-25T17:40:53Z) - SIR: Self-supervised Image Rectification via Seeing the Same Scene from
Multiple Different Lenses [82.56853587380168]
We propose a novel self-supervised image rectification (SIR) method based on an important insight that the rectified results of distorted images of the same scene from different lens should be the same.
We leverage a differentiable warping module to generate the rectified images and re-distorted images from the distortion parameters.
Our method achieves comparable or even better performance than the supervised baseline method and representative state-of-the-art methods.
arXiv Detail & Related papers (2020-11-30T08:23:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.