Deep Reinforcement Learning Using a Low-Dimensional Observation Filter
for Visual Complex Video Game Playing
- URL: http://arxiv.org/abs/2204.11370v1
- Date: Sun, 24 Apr 2022 22:17:08 GMT
- Title: Deep Reinforcement Learning Using a Low-Dimensional Observation Filter
for Visual Complex Video Game Playing
- Authors: Victor Augusto Kich, Junior Costa de Jesus, Ricardo Bedin Grando,
Alisson Henrique Kolling, Gabriel Vin\'icius Heisler, Rodrigo da Silva Guerra
- Abstract summary: It requires the processing of large amounts of data from high-dimensional observation spaces, frame by frame, and the agent's actions are computed according to deep neural network policies.
In this paper, we propose a low-dimensional observation filter that allows a deep Q-network agent to successfully play in a visually complex and modern video-game, called Neon Drive.
- Score: 1.2468700211588883
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep Reinforcement Learning (DRL) has produced great achievements since it
was proposed, including the possibility of processing raw vision input data.
However, training an agent to perform tasks based on image feedback remains a
challenge. It requires the processing of large amounts of data from
high-dimensional observation spaces, frame by frame, and the agent's actions
are computed according to deep neural network policies, end-to-end. Image
pre-processing is an effective way of reducing these high dimensional spaces,
eliminating unnecessary information present in the scene, supporting the
extraction of features and their representations in the agent's neural network.
Modern video-games are examples of this type of challenge for DRL algorithms
because of their visual complexity. In this paper, we propose a low-dimensional
observation filter that allows a deep Q-network agent to successfully play in a
visually complex and modern video-game, called Neon Drive.
Related papers
- An Examination of Offline-Trained Encoders in Vision-Based Deep Reinforcement Learning for Autonomous Driving [0.0]
Research investigates the challenges Deep Reinforcement Learning (DRL) faces in Partially Observable Markov Decision Processes (POMDP)
Our research adopts an offline-trained encoder to leverage large video datasets through self-supervised learning to learn generalizable representations.
We show that the features learned by watching BDD100K driving videos can be directly transferred to achieve lane following and collision avoidance in CARLA simulator.
arXiv Detail & Related papers (2024-09-02T14:16:23Z) - Neural Radiance Field Codebooks [53.01356339021285]
We introduce Neural Radiance Field Codebooks (NRC), a scalable method for learning object-centric representations.
NRC learns to reconstruct scenes from novel views using a dictionary of object codes which are decoded through a volumetric reconstruction.
We show that NRC representations transfer well to object navigation in THOR, outperforming 2D and 3D representation learning methods by 3.1% success rate.
arXiv Detail & Related papers (2023-01-10T18:03:48Z) - An Empirical Study of Remote Sensing Pretraining [117.90699699469639]
We conduct an empirical study of remote sensing pretraining (RSP) on aerial images.
RSP can help deliver distinctive performances in scene recognition tasks.
RSP mitigates the data discrepancies of traditional ImageNet pretraining on RS images, but it may still suffer from task discrepancies.
arXiv Detail & Related papers (2022-04-06T13:38:11Z) - A Deeper Look into DeepCap [96.67706102518238]
We propose a novel deep learning approach for monocular dense human performance capture.
Our method is trained in a weakly supervised manner based on multi-view supervision.
Our approach outperforms the state of the art in terms of quality and robustness.
arXiv Detail & Related papers (2021-11-20T11:34:33Z) - Deep Direct Volume Rendering: Learning Visual Feature Mappings From
Exemplary Images [57.253447453301796]
We introduce Deep Direct Volume Rendering (DeepDVR), a generalization of Direct Volume Rendering (DVR) that allows for the integration of deep neural networks into the DVR algorithm.
We conceptualize the rendering in a latent color space, thus enabling the use of deep architectures to learn implicit mappings for feature extraction and classification.
Our generalization serves to derive novel volume rendering architectures that can be trained end-to-end directly from examples in image space.
arXiv Detail & Related papers (2021-06-09T23:03:00Z) - Improving Computational Efficiency in Visual Reinforcement Learning via
Stored Embeddings [89.63764845984076]
We present Stored Embeddings for Efficient Reinforcement Learning (SEER)
SEER is a simple modification of existing off-policy deep reinforcement learning methods.
We show that SEER does not degrade the performance of RLizable agents while significantly saving computation and memory.
arXiv Detail & Related papers (2021-03-04T08:14:10Z) - Are Gradient-based Saliency Maps Useful in Deep Reinforcement Learning? [4.254099382808598]
Deep Reinforcement Learning (DRL) connects the classic Reinforcement Learning algorithms with Deep Neural Networks.
A problem in DRL is that CNNs are black-boxes and it is hard to understand the decision-making process of agents.
This work brings some of the best-known visualization methods from the field of image classification to the area of Deep Reinforcement Learning.
arXiv Detail & Related papers (2020-12-02T15:38:36Z) - Accurate RGB-D Salient Object Detection via Collaborative Learning [101.82654054191443]
RGB-D saliency detection shows impressive ability on some challenge scenarios.
We propose a novel collaborative learning framework where edge, depth and saliency are leveraged in a more efficient way.
arXiv Detail & Related papers (2020-07-23T04:33:36Z) - Top-Down Networks: A coarse-to-fine reimagination of CNNs [25.079310083166824]
Biological vision adopts a coarse-to-fine information processing pathway.
Top-down networks offer a line of defence against adversarial attacks that introduce high frequency noise.
This paper offers empirical evidence for the applicability of the top-down resolution processing to various existing architectures on multiple visual tasks.
arXiv Detail & Related papers (2020-04-16T12:29:48Z) - Acceleration of Actor-Critic Deep Reinforcement Learning for Visual
Grasping in Clutter by State Representation Learning Based on Disentanglement
of a Raw Input Image [4.970364068620608]
Actor-critic deep reinforcement learning (RL) methods typically perform very poorly when grasping diverse objects.
We employ state representation learning (SRL), where we encode essential information first for subsequent use in RL.
We found that preprocessing based on the disentanglement of a raw input image is the key to effectively capturing a compact representation.
arXiv Detail & Related papers (2020-02-27T03:58:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.