Visual Radial Basis Q-Network
- URL: http://arxiv.org/abs/2206.06712v1
- Date: Tue, 14 Jun 2022 09:34:34 GMT
- Title: Visual Radial Basis Q-Network
- Authors: Julien Hautot, C\'eline Teuliere and Nourddine Azzaoui
- Abstract summary: We propose a generic method to extract sparse features from raw images with few trainable parameters.
We show that the proposed approach provides similar or, in some cases, even better performances with fewer trainable parameters while being conceptually simpler.
- Score: 0.2148535041822524
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While reinforcement learning (RL) from raw images has been largely
investigated in the last decade, existing approaches still suffer from a number
of constraints. The high input dimension is often handled using either expert
knowledge to extract handcrafted features or environment encoding through
convolutional networks. Both solutions require numerous parameters to be
optimized. In contrast, we propose a generic method to extract sparse features
from raw images with few trainable parameters. We achieved this using a Radial
Basis Function Network (RBFN) directly on raw image. We evaluate the
performance of the proposed approach for visual extraction in Q-learning tasks
in the Vizdoom environment. Then, we compare our results with two Deep
Q-Network, one trained directly on images and another one trained on feature
extracted by a pretrained auto-encoder. We show that the proposed approach
provides similar or, in some cases, even better performances with fewer
trainable parameters while being conceptually simpler.
Related papers
- CricaVPR: Cross-image Correlation-aware Representation Learning for Visual Place Recognition [73.51329037954866]
We propose a robust global representation method with cross-image correlation awareness for visual place recognition.
Our method uses the attention mechanism to correlate multiple images within a batch.
Our method outperforms state-of-the-art methods by a large margin with significantly less training time.
arXiv Detail & Related papers (2024-02-29T15:05:11Z) - Hyper-VolTran: Fast and Generalizable One-Shot Image to 3D Object
Structure via HyperNetworks [53.67497327319569]
We introduce a novel neural rendering technique to solve image-to-3D from a single view.
Our approach employs the signed distance function as the surface representation and incorporates generalizable priors through geometry-encoding volumes and HyperNetworks.
Our experiments show the advantages of our proposed approach with consistent results and rapid generation.
arXiv Detail & Related papers (2023-12-24T08:42:37Z) - Reinforcement Learning from Diffusion Feedback: Q* for Image Search [2.5835347022640254]
We present two models for image generation using model-agnostic learning.
RLDF is a singular approach for visual imitation through prior-preserving reward function guidance.
It generates high-quality images over varied domains showcasing class-consistency and strong visual diversity.
arXiv Detail & Related papers (2023-11-27T09:20:12Z) - Raw or Cooked? Object Detection on RAW Images [11.991240159496833]
We investigate the hypothesis that the intermediate representation of visually pleasing images is sub-optimal for downstream computer vision tasks.
We suggest that the operations of the ISP instead should be optimized towards the end task, by learning the parameters of the operations jointly during training.
We propose a new learnable operation that enables an object detector to achieve superior performance when compared to both previous works and traditional RGB images.
arXiv Detail & Related papers (2023-01-21T15:42:53Z) - Rank-Enhanced Low-Dimensional Convolution Set for Hyperspectral Image
Denoising [50.039949798156826]
This paper tackles the challenging problem of hyperspectral (HS) image denoising.
We propose rank-enhanced low-dimensional convolution set (Re-ConvSet)
We then incorporate Re-ConvSet into the widely-used U-Net architecture to construct an HS image denoising method.
arXiv Detail & Related papers (2022-07-09T13:35:12Z) - Image Restoration by Deep Projected GSURE [115.57142046076164]
Ill-posed inverse problems appear in many image processing applications, such as deblurring and super-resolution.
We propose a new image restoration framework that is based on minimizing a loss function that includes a "projected-version" of the Generalized SteinUnbiased Risk Estimator (GSURE) and parameterization of the latent image by a CNN.
arXiv Detail & Related papers (2021-02-04T08:52:46Z) - Deep Iterative Residual Convolutional Network for Single Image
Super-Resolution [31.934084942626257]
We propose a deep Iterative Super-Resolution Residual Convolutional Network (ISRResCNet)
It exploits the powerful image regularization and large-scale optimization techniques by training the deep network in an iterative manner with a residual learning approach.
Our method with a few trainable parameters improves the results for different scaling factors in comparison with the state-of-art methods.
arXiv Detail & Related papers (2020-09-07T12:54:14Z) - A deep primal-dual proximal network for image restoration [8.797434238081372]
We design a deep network, named DeepPDNet, built from primal-dual iterations associated with the minimization of a standard penalized likelihood with an analysis prior.
Two different learning strategies: "Full learning" and "Partial learning" are proposed, the first one is the most efficient numerically.
Extensive results show that the proposed DeepPDNet demonstrates excellent performance on the MNIST and the more complex BSD68, BSD100, and SET14 datasets for image restoration and single image super-resolution task.
arXiv Detail & Related papers (2020-07-02T08:29:52Z) - The Power of Triply Complementary Priors for Image Compressive Sensing [89.14144796591685]
We propose a joint low-rank deep (LRD) image model, which contains a pair of complementaryly trip priors.
We then propose a novel hybrid plug-and-play framework based on the LRD model for image CS.
To make the optimization tractable, a simple yet effective algorithm is proposed to solve the proposed H-based image CS problem.
arXiv Detail & Related papers (2020-05-16T08:17:44Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.