Delta Schema Network in Model-based Reinforcement Learning
- URL: http://arxiv.org/abs/2006.09950v2
- Date: Wed, 8 Jul 2020 05:58:54 GMT
- Title: Delta Schema Network in Model-based Reinforcement Learning
- Authors: Andrey Gorodetskiy, Alexandra Shlychkova, Aleksandr I. Panov
- Abstract summary: This work is devoted to unresolved problems of Artificial General Intelligence - the inefficiency of transfer learning.
We are expanding the schema networks method which allows to extract the logical relationships between objects and actions from the environment data.
We present algorithms for training a Delta Network (DSN), predicting future states of the environment and planning actions that will lead to positive reward.
- Score: 125.99533416395765
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work is devoted to unresolved problems of Artificial General
Intelligence - the inefficiency of transfer learning. One of the mechanisms
that are used to solve this problem in the area of reinforcement learning is a
model-based approach. In the paper we are expanding the schema networks method
which allows to extract the logical relationships between objects and actions
from the environment data. We present algorithms for training a Delta Schema
Network (DSN), predicting future states of the environment and planning actions
that will lead to positive reward. DSN shows strong performance of transfer
learning on the classic Atari game environment.
Related papers
- Transfer Learning Under High-Dimensional Network Convolutional Regression Model [20.18595334666282]
We propose a high-dimensional transfer learning framework based on network convolutional regression ( NCR)
Our methodology includes a two-step transfer learning algorithm that addresses domain shift between source and target networks.
Empirical evaluations, including simulations and a real-world application using Sina Weibo data, demonstrate substantial improvements in prediction accuracy.
arXiv Detail & Related papers (2025-04-28T16:52:28Z) - Latent Diffusion Planning for Imitation Learning [78.56207566743154]
Latent Diffusion Planning (LDP) is a modular approach consisting of a planner and inverse dynamics model.
By separating planning from action prediction, LDP can benefit from the denser supervision signals of suboptimal and action-free data.
On simulated visual robotic manipulation tasks, LDP outperforms state-of-the-art imitation learning approaches.
arXiv Detail & Related papers (2025-04-23T17:53:34Z) - Robo-taxi Fleet Coordination at Scale via Reinforcement Learning [21.266509380044912]
This work introduces a novel decision-making framework that unites mathematical modeling with data-driven techniques.
We present the AMoD coordination problem through the lens of reinforcement learning and propose a graph network-based framework.
In particular, we present the AMoD coordination problem through the lens of reinforcement learning and propose a graph network-based framework.
arXiv Detail & Related papers (2025-04-08T15:19:41Z) - Contrastive Representation Learning for Dynamic Link Prediction in Temporal Networks [1.9389881806157312]
We introduce a self-supervised method for learning representations of temporal networks.
We propose a recurrent message-passing neural network architecture for modeling the information flow over time-respecting paths of temporal networks.
The proposed method is tested on Enron, COLAB, and Facebook datasets.
arXiv Detail & Related papers (2024-08-22T22:50:46Z) - Dynamic Encoding and Decoding of Information for Split Learning in
Mobile-Edge Computing: Leveraging Information Bottleneck Theory [1.1151919978983582]
Split learning is a privacy-preserving distributed learning paradigm in which an ML model is split into two parts (i.e., an encoder and a decoder)
In mobile-edge computing, network functions can be trained via split learning where an encoder resides in a user equipment (UE) and a decoder resides in the edge network.
We present a new framework and training mechanism to enable a dynamic balancing of the transmission resource consumption with the informativeness of the shared latent representations.
arXiv Detail & Related papers (2023-09-06T07:04:37Z) - Common Knowledge Learning for Generating Transferable Adversarial
Examples [60.1287733223249]
This paper focuses on an important type of black-box attacks, where the adversary generates adversarial examples by a substitute (source) model.
Existing methods tend to give unsatisfactory adversarial transferability when the source and target models are from different types of DNN architectures.
We propose a common knowledge learning (CKL) framework to learn better network weights to generate adversarial examples.
arXiv Detail & Related papers (2023-07-01T09:07:12Z) - Towards a Better Theoretical Understanding of Independent Subnetwork Training [56.24689348875711]
We take a closer theoretical look at Independent Subnetwork Training (IST)
IST is a recently proposed and highly effective technique for solving the aforementioned problems.
We identify fundamental differences between IST and alternative approaches, such as distributed methods with compressed communication.
arXiv Detail & Related papers (2023-06-28T18:14:22Z) - Model-Based Machine Learning for Communications [110.47840878388453]
We review existing strategies for combining model-based algorithms and machine learning from a high level perspective.
We focus on symbol detection, which is one of the fundamental tasks of communication receivers.
arXiv Detail & Related papers (2021-01-12T19:55:34Z) - An Ode to an ODE [78.97367880223254]
We present a new paradigm for Neural ODE algorithms, called ODEtoODE, where time-dependent parameters of the main flow evolve according to a matrix flow on the group O(d)
This nested system of two flows provides stability and effectiveness of training and provably solves the gradient vanishing-explosion problem.
arXiv Detail & Related papers (2020-06-19T22:05:19Z) - Deep learning of contagion dynamics on complex networks [0.0]
We propose a complementary approach based on deep learning to build effective models of contagion dynamics on networks.
By allowing simulations on arbitrary network structures, our approach makes it possible to explore the properties of the learned dynamics beyond the training data.
Our results demonstrate how deep learning offers a new and complementary perspective to build effective models of contagion dynamics on networks.
arXiv Detail & Related papers (2020-06-09T17:18:34Z) - Deep Learning of Dynamic Subsurface Flow via Theory-guided Generative
Adversarial Network [0.0]
Theory-guided generative adversarial network (TgGAN) is proposed to solve dynamic partial differential equations (PDEs)
TgGAN is proposed for dynamic subsurface flow with heterogeneous model parameters.
Numerical results demonstrate that the TgGAN model is robust and reliable for deep learning of dynamic PDEs.
arXiv Detail & Related papers (2020-06-02T02:53:26Z) - Deep Learning for Ultra-Reliable and Low-Latency Communications in 6G
Networks [84.2155885234293]
We first summarize how to apply data-driven supervised deep learning and deep reinforcement learning in URLLC.
To address these open problems, we develop a multi-level architecture that enables device intelligence, edge intelligence, and cloud intelligence for URLLC.
arXiv Detail & Related papers (2020-02-22T14:38:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.