Q-SMASH: Q-Learning-based Self-Adaptation of Human-Centered Internet of
Things
- URL: http://arxiv.org/abs/2107.05949v1
- Date: Tue, 13 Jul 2021 09:41:05 GMT
- Title: Q-SMASH: Q-Learning-based Self-Adaptation of Human-Centered Internet of
Things
- Authors: Hamed Rahimi, Iago Felipe Trentin, Fano Ramparany, Olivier Boissier
- Abstract summary: This article presents Q-SMASH, a reinforcement learning-based approach for self-adaptation of IoT objects in human-centered environments.
Q-SMASH aims to learn the behaviors of users along with respecting human values.
The learning ability of Q-SMASH allows it to adapt itself to the behavioral change of users and make more accurate decisions.
- Score: 0.8602553195689512
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As the number of Human-Centered Internet of Things (HCIoT) applications
increases, the self-adaptation of its services and devices is becoming a
fundamental requirement for addressing the uncertainties of the environment in
decision-making processes. Self-adaptation of HCIoT aims to manage run-time
changes in a dynamic environment and to adjust the functionality of IoT objects
in order to achieve desired goals during execution. SMASH is a semantic-enabled
multi-agent system for self-adaptation of HCIoT that autonomously adapts IoT
objects to uncertainties of their environment. SMASH addresses the
self-adaptation of IoT applications only according to the human values of
users, while the behavior of users is not addressed. This article presents
Q-SMASH: a multi-agent reinforcement learning-based approach for
self-adaptation of IoT objects in human-centered environments. Q-SMASH aims to
learn the behaviors of users along with respecting human values. The learning
ability of Q-SMASH allows it to adapt itself to the behavioral change of users
and make more accurate decisions in different states and situations.
Related papers
- Metacognition for Unknown Situations and Environments (MUSE) [3.2020845462590697]
We propose the Metacognition for Unknown Situations and Environments (MUSE) framework.
MUSE integrates metacognitive processes--specifically self-awareness and self-regulation--into autonomous agents.
Agents show significant improvements in self-awareness and self-regulation.
arXiv Detail & Related papers (2024-11-20T18:41:03Z) - HAZARD Challenge: Embodied Decision Making in Dynamically Changing
Environments [93.94020724735199]
HAZARD consists of three unexpected disaster scenarios, including fire, flood, and wind.
This benchmark enables us to evaluate autonomous agents' decision-making capabilities across various pipelines.
arXiv Detail & Related papers (2024-01-23T18:59:43Z) - The Internet of Senses: Building on Semantic Communications and Edge
Intelligence [67.75406096878321]
The Internet of Senses (IoS) holds the promise of flawless telepresence-style communication for all human receptors'
We elaborate on how the emerging semantic communications and Artificial Intelligence (AI)/Machine Learning (ML) paradigms may satisfy the requirements of IoS use cases.
arXiv Detail & Related papers (2022-12-21T03:37:38Z) - Goal-Conditioned Q-Learning as Knowledge Distillation [136.79415677706612]
We explore a connection between off-policy reinforcement learning in goal-conditioned settings and knowledge distillation.
We empirically show that this can improve the performance of goal-conditioned off-policy reinforcement learning when the space of goals is high-dimensional.
We also show that this technique can be adapted to allow for efficient learning in the case of multiple simultaneous sparse goals.
arXiv Detail & Related papers (2022-08-28T22:01:10Z) - Autonomous Open-Ended Learning of Tasks with Non-Stationary
Interdependencies [64.0476282000118]
Intrinsic motivations have proven to generate a task-agnostic signal to properly allocate the training time amongst goals.
While the majority of works in the field of intrinsically motivated open-ended learning focus on scenarios where goals are independent from each other, only few of them studied the autonomous acquisition of interdependent tasks.
In particular, we first deepen the analysis of a previous system, showing the importance of incorporating information about the relationships between tasks at a higher level of the architecture.
Then we introduce H-GRAIL, a new system that extends the previous one by adding a new learning layer to store the autonomously acquired sequences
arXiv Detail & Related papers (2022-05-16T10:43:01Z) - Learning to Walk Autonomously via Reset-Free Quality-Diversity [73.08073762433376]
Quality-Diversity algorithms can discover large and complex behavioural repertoires consisting of both diverse and high-performing skills.
Existing QD algorithms need large numbers of evaluations as well as episodic resets, which require manual human supervision and interventions.
This paper proposes Reset-Free Quality-Diversity optimization (RF-QD) as a step towards autonomous learning for robotics in open-ended environments.
arXiv Detail & Related papers (2022-04-07T14:07:51Z) - Backprop-Free Reinforcement Learning with Active Neural Generative
Coding [84.11376568625353]
We propose a computational framework for learning action-driven generative models without backpropagation of errors (backprop) in dynamic environments.
We develop an intelligent agent that operates even with sparse rewards, drawing inspiration from the cognitive theory of planning as inference.
The robust performance of our agent offers promising evidence that a backprop-free approach for neural inference and learning can drive goal-directed behavior.
arXiv Detail & Related papers (2021-07-10T19:02:27Z) - SMASH: a Semantic-enabled Multi-agent Approach for Self-adaptation of
Human-centered IoT [0.8602553195689512]
This paper presents SMASH: a multi-agent approach for self-adaptation of IoT applications in human-centered environments.
SMASH agents are provided with a 4-layer architecture based on the BDI agent model that integrates human values with goal-reasoning, planning, and acting.
arXiv Detail & Related papers (2021-05-31T12:33:27Z) - FaiR-IoT: Fairness-aware Human-in-the-Loop Reinforcement Learning for
Harnessing Human Variability in Personalized IoT [0.0]
FaiR-IoT is a reinforcement learning-based framework for adaptive and fairness-aware human-in-the-loop IoT applications.
We validate the proposed framework on two applications, namely (i) Human-in-the-Loop Automotive Advanced Driver Assistance Systems and (ii) Human-in-the-Loop Smart House.
Results obtained on these two applications validate the generality of FaiR-IoT and its ability to provide a personalized experience.
arXiv Detail & Related papers (2021-03-30T02:30:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.