Towards Multi-Agent Reinforcement Learning using Quantum Boltzmann
Machines
- URL: http://arxiv.org/abs/2109.10900v1
- Date: Wed, 22 Sep 2021 17:59:24 GMT
- Title: Towards Multi-Agent Reinforcement Learning using Quantum Boltzmann
Machines
- Authors: Tobias M\"uller, Christoph Roch, Kyrill Schmid and Philipp Altmann
- Abstract summary: We propose an extension to the original concept in order to solve more challenging problems.
We add an experience replay buffer and use different networks for approximating the target and policy values.
Quantum sampling proves to be a promising method for reinforcement learning tasks, but is currently limited by the QPU size.
- Score: 2.015864965523243
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Reinforcement learning has driven impressive advances in machine learning.
Simultaneously, quantum-enhanced machine learning algorithms using quantum
annealing underlie heavy developments. Recently, a multi-agent reinforcement
learning (MARL) architecture combining both paradigms has been proposed. This
novel algorithm, which utilizes Quantum Boltzmann Machines (QBMs) for Q-value
approximation has outperformed regular deep reinforcement learning in terms of
time-steps needed to converge. However, this algorithm was restricted to
single-agent and small 2x2 multi-agent grid domains. In this work, we propose
an extension to the original concept in order to solve more challenging
problems. Similar to classic DQNs, we add an experience replay buffer and use
different networks for approximating the target and policy values. The
experimental results show that learning becomes more stable and enables agents
to find optimal policies in grid-domains with higher complexity. Additionally,
we assess how parameter sharing influences the agents behavior in multi-agent
domains. Quantum sampling proves to be a promising method for reinforcement
learning tasks, but is currently limited by the QPU size and therefore by the
size of the input and Boltzmann machine.
Related papers
- A Matrix Product State Model for Simultaneous Classification and Generation [0.8192907805418583]
Quantum machine learning (QML) is a rapidly expanding field that merges the principles of quantum computing with the techniques of machine learning.
Here, we present a novel matrix product state (MPS) model, where the MPS functions as both a classifier and a generator.
Our contributions offer insights into the mechanics of tensor network methods for generation tasks.
arXiv Detail & Related papers (2024-06-25T10:23:36Z) - Quantum Multi-Agent Reinforcement Learning for Aerial Ad-hoc Networks [0.19791587637442667]
This paper presents an aerial communication use case and introduces a hybrid quantum-classical (HQC) ML algorithm to solve it.
Results show a slight increase in performance for the quantum-enhanced solution with respect to a comparable classical algorithm.
These promising results show the potential of QMARL to industrially-relevant complex use cases.
arXiv Detail & Related papers (2024-04-26T15:57:06Z) - Quantum Subroutine for Variance Estimation: Algorithmic Design and Applications [80.04533958880862]
Quantum computing sets the foundation for new ways of designing algorithms.
New challenges arise concerning which field quantum speedup can be achieved.
Looking for the design of quantum subroutines that are more efficient than their classical counterpart poses solid pillars to new powerful quantum algorithms.
arXiv Detail & Related papers (2024-02-26T09:32:07Z) - Multi-Agent Quantum Reinforcement Learning using Evolutionary
Optimization [7.305065320738301]
We build upon an existing approach for gradient free Quantum Reinforcement Learning and propose three genetic variations with Variational Quantum Circuits for Multi-Agent Reinforcement Learning.
We show that our Variational Quantum Circuit approaches perform significantly better compared to a neural network with a similar amount of trainable parameters.
arXiv Detail & Related papers (2023-11-09T17:45:32Z) - Inverse Factorized Q-Learning for Cooperative Multi-agent Imitation
Learning [13.060023718506917]
imitation learning (IL) is a problem of learning to mimic expert behaviors from demonstrations in cooperative multi-agent systems.
We introduce a novel multi-agent IL algorithm designed to address these challenges.
Our approach enables the centralized learning by leveraging mixing networks to aggregate decentralized Q functions.
arXiv Detail & Related papers (2023-10-10T17:11:20Z) - Optimizing Tensor Network Contraction Using Reinforcement Learning [86.05566365115729]
We propose a Reinforcement Learning (RL) approach combined with Graph Neural Networks (GNN) to address the contraction ordering problem.
The problem is extremely challenging due to the huge search space, the heavy-tailed reward distribution, and the challenging credit assignment.
We show how a carefully implemented RL-agent that uses a GNN as the basic policy construct can address these challenges.
arXiv Detail & Related papers (2022-04-18T21:45:13Z) - MQBench: Towards Reproducible and Deployable Model Quantization
Benchmark [53.12623958951738]
MQBench is a first attempt to evaluate, analyze, and benchmark the and deployability for model quantization algorithms.
We choose multiple platforms for real-world deployments, including CPU, GPU, ASIC, DSP, and evaluate extensive state-of-the-art quantization algorithms.
We conduct a comprehensive analysis and find considerable intuitive or counter-intuitive insights.
arXiv Detail & Related papers (2021-11-05T23:38:44Z) - Quantum Reinforcement Learning: the Maze problem [11.240669509034298]
We will introduce a new QML model generalizing the classical concept of Reinforcement Learning to the quantum domain.
In particular, we apply this idea to the maze problem, where an agent has to learn the optimal set of actions in order to escape from a maze with the highest success probability.
We find that the agent learns the optimal strategy in both the classical and quantum regimes, and we also investigate its behaviour in a noisy environment.
arXiv Detail & Related papers (2021-08-10T07:49:58Z) - Efficient Model-Based Multi-Agent Mean-Field Reinforcement Learning [89.31889875864599]
We propose an efficient model-based reinforcement learning algorithm for learning in multi-agent systems.
Our main theoretical contributions are the first general regret bounds for model-based reinforcement learning for MFC.
We provide a practical parametrization of the core optimization problem.
arXiv Detail & Related papers (2021-07-08T18:01:02Z) - UPDeT: Universal Multi-agent Reinforcement Learning via Policy
Decoupling with Transformers [108.92194081987967]
We make the first attempt to explore a universal multi-agent reinforcement learning pipeline, designing one single architecture to fit tasks.
Unlike previous RNN-based models, we utilize a transformer-based model to generate a flexible policy.
The proposed model, named as Universal Policy Decoupling Transformer (UPDeT), further relaxes the action restriction and makes the multi-agent task's decision process more explainable.
arXiv Detail & Related papers (2021-01-20T07:24:24Z) - SUNRISE: A Simple Unified Framework for Ensemble Learning in Deep
Reinforcement Learning [102.78958681141577]
We present SUNRISE, a simple unified ensemble method, which is compatible with various off-policy deep reinforcement learning algorithms.
SUNRISE integrates two key ingredients: (a) ensemble-based weighted Bellman backups, which re-weight target Q-values based on uncertainty estimates from a Q-ensemble, and (b) an inference method that selects actions using the highest upper-confidence bounds for efficient exploration.
arXiv Detail & Related papers (2020-07-09T17:08:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.