Self-Gated Memory Recurrent Network for Efficient Scalable HDR
Deghosting
- URL: http://arxiv.org/abs/2112.13050v1
- Date: Fri, 24 Dec 2021 12:36:33 GMT
- Title: Self-Gated Memory Recurrent Network for Efficient Scalable HDR
Deghosting
- Authors: K. Ram Prabhakar, Susmit Agrawal, R. Venkatesh Babu
- Abstract summary: We propose a novel recurrent network-based HDR deghosting method for fusing arbitrary length dynamic sequences.
We introduce a new recurrent cell architecture, namely Self-Gated Memory (SGM) cell, that outperforms the standard LSTM cell.
The proposed approach achieves state-of-the-art performance compared to existing HDR deghosting methods quantitatively across three publicly available datasets.
- Score: 59.04604001936661
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a novel recurrent network-based HDR deghosting method for fusing
arbitrary length dynamic sequences. The proposed method uses convolutional and
recurrent architectures to generate visually pleasing, ghosting-free HDR
images. We introduce a new recurrent cell architecture, namely Self-Gated
Memory (SGM) cell, that outperforms the standard LSTM cell while containing
fewer parameters and having faster running times. In the SGM cell, the
information flow through a gate is controlled by multiplying the gate's output
by a function of itself. Additionally, we use two SGM cells in a bidirectional
setting to improve output quality. The proposed approach achieves
state-of-the-art performance compared to existing HDR deghosting methods
quantitatively across three publicly available datasets while simultaneously
achieving scalability to fuse variable-length input sequence without
necessitating re-training. Through extensive ablations, we demonstrate the
importance of individual components in our proposed approach. The code is
available at https://val.cds.iisc.ac.in/HDR/HDRRNN/index.html.
Related papers
- SelfDRSC++: Self-Supervised Learning for Dual Reversed Rolling Shutter Correction [72.05587640928879]
We propose an enhanced Self-supervised learning framework for Dual reversed RS distortion Correction (SelfDRSC++)
We introduce a lightweight DRSC network that incorporates a bidirectional correlation matching block to refine the joint optimization of optical flows and corrected RS features.
To effectively train the DRSC network, we propose a self-supervised learning strategy that ensures cycle consistency between input and reconstructed dual reversed RS images.
arXiv Detail & Related papers (2024-08-21T08:17:22Z) - MF-NeRF: Memory Efficient NeRF with Mixed-Feature Hash Table [62.164549651134465]
We propose MF-NeRF, a memory-efficient NeRF framework that employs a Mixed-Feature hash table to improve memory efficiency and reduce training time while maintaining reconstruction quality.
Our experiments with state-of-the-art Instant-NGP, TensoRF, and DVGO, indicate our MF-NeRF could achieve the fastest training time on the same GPU hardware with similar or even higher reconstruction quality.
arXiv Detail & Related papers (2023-04-25T05:44:50Z) - Raising The Limit Of Image Rescaling Using Auxiliary Encoding [7.9700865143145485]
Recently, image rescaling models like IRN utilize the bidirectional nature of INN to push the performance limit of image upscaling.
We propose auxiliary encoding modules to further push the limit of image rescaling performance.
arXiv Detail & Related papers (2023-03-12T20:49:07Z) - Super-Resolution Neural Operator [5.018040244860608]
We propose a framework that can resolve high-resolution (HR) images at arbitrary scales from the low-resolution (LR) counterparts.
Treating the LR-HR image pairs as continuous functions approximated with different grid sizes, SRNO learns the mapping between the corresponding function spaces.
Experiments show that SRNO outperforms existing continuous SR methods in terms of both accuracy and running time.
arXiv Detail & Related papers (2023-03-05T06:17:43Z) - GLEAM: Greedy Learning for Large-Scale Accelerated MRI Reconstruction [50.248694764703714]
Unrolled neural networks have recently achieved state-of-the-art accelerated MRI reconstruction.
These networks unroll iterative optimization algorithms by alternating between physics-based consistency and neural-network based regularization.
We propose Greedy LEarning for Accelerated MRI reconstruction, an efficient training strategy for high-dimensional imaging settings.
arXiv Detail & Related papers (2022-07-18T06:01:29Z) - Hypercomplex Image-to-Image Translation [13.483068375377362]
Image-to-image translation (I2I) aims at transferring the content representation from an input domain to an output one.
Recent I2I generative models, which gain outstanding results in this task, comprise a set of diverse deep networks each with tens of million parameters.
We propose to leverage hypercomplex algebra properties to define lightweight I2I generative models capable of preserving pre-existing relations among image dimensions.
arXiv Detail & Related papers (2022-05-04T14:28:50Z) - Efficient Re-parameterization Residual Attention Network For
Nonhomogeneous Image Dehazing [4.723586858098229]
ERRA-Net has an impressive speed, processing 1200x1600 HD quality images with an average runtime of 166.11 fps.
We use cascaded MA blocks to extract high-frequency features step by step, and the Multi-layer attention fusion tail combines the shallow and deep features of the model to get the residual of the clean image.
arXiv Detail & Related papers (2021-09-12T10:03:44Z) - MFGNet: Dynamic Modality-Aware Filter Generation for RGB-T Tracking [72.65494220685525]
We propose a new dynamic modality-aware filter generation module (named MFGNet) to boost the message communication between visible and thermal data.
We generate dynamic modality-aware filters with two independent networks. The visible and thermal filters will be used to conduct a dynamic convolutional operation on their corresponding input feature maps respectively.
To address issues caused by heavy occlusion, fast motion, and out-of-view, we propose to conduct a joint local and global search by exploiting a new direction-aware target-driven attention mechanism.
arXiv Detail & Related papers (2021-07-22T03:10:51Z) - Efficient Two-Stream Network for Violence Detection Using Separable
Convolutional LSTM [0.0]
We propose an efficient two-stream deep learning architecture leveraging Separable Convolutional LSTM (SepConvLSTM) and pre-trained MobileNet.
SepConvLSTM is constructed by replacing convolution operation at each gate of ConvLSTM with a depthwise separable convolution.
Our model outperforms the accuracy on the larger and more challenging RWF-2000 dataset by more than a 2% margin.
arXiv Detail & Related papers (2021-02-21T12:01:48Z) - Self Sparse Generative Adversarial Networks [73.590634413751]
Generative Adversarial Networks (GANs) are an unsupervised generative model that learns data distribution through adversarial training.
We propose a Self Sparse Generative Adversarial Network (Self-Sparse GAN) that reduces the parameter space and alleviates the zero gradient problem.
arXiv Detail & Related papers (2021-01-26T04:49:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.