Data-driven Koopman Operators for Model-based Shared Control of
Human-Machine Systems
- URL: http://arxiv.org/abs/2006.07210v1
- Date: Fri, 12 Jun 2020 14:14:07 GMT
- Title: Data-driven Koopman Operators for Model-based Shared Control of
Human-Machine Systems
- Authors: Alexander Broad, Ian Abraham, Todd Murphey, Brenna Argall
- Abstract summary: We present a data-driven shared control algorithm that can be used to improve a human operator's control of complex machines.
Both the dynamics and information about the user's interaction are learned from observation through the use of a Koopman operator.
We find that model-based shared control significantly improves task and control metrics when compared to a natural learning, or user only, control paradigm.
- Score: 66.65503164312705
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a data-driven shared control algorithm that can be used to improve
a human operator's control of complex dynamic machines and achieve tasks that
would otherwise be challenging, or impossible, for the user on their own. Our
method assumes no a priori knowledge of the system dynamics. Instead, both the
dynamics and information about the user's interaction are learned from
observation through the use of a Koopman operator. Using the learned model, we
define an optimization problem to compute the autonomous partner's control
policy. Finally, we dynamically allocate control authority to each partner
based on a comparison of the user input and the autonomously generated control.
We refer to this idea as model-based shared control (MbSC). We evaluate the
efficacy of our approach with two human subjects studies consisting of 32 total
participants (16 subjects in each study). The first study imposes a linear
constraint on the modeling and autonomous policy generation algorithms. The
second study explores the more general, nonlinear variant. Overall, we find
that model-based shared control significantly improves task and control metrics
when compared to a natural learning, or user only, control paradigm. Our
experiments suggest that models learned via the Koopman operator generalize
across users, indicating that it is not necessary to collect data from each
individual user before providing assistance with MbSC. We also demonstrate the
data-efficiency of MbSC and consequently, it's usefulness in online learning
paradigms. Finally, we find that the nonlinear variant has a greater impact on
a user's ability to successfully achieve a defined task than the linear
variant.
Related papers
- Nonparametric Control-Koopman Operator Learning: Flexible and Scalable Models for Prediction and Control [2.7784144651669704]
We present a nonparametric framework for learning Koopman operator representations of nonlinear control-affine systems.
We also enhance the scalability of control-Koopman operator estimators by leveraging random projections.
The efficacy of our novel cKOR approach is demonstrated on both forecasting and control tasks.
arXiv Detail & Related papers (2024-05-12T15:46:52Z) - Physics-informed reinforcement learning via probabilistic co-adjustment
functions [3.6787556334630334]
We introduce co-kriging adjustments (CKA) and ridge regression adjustment (RRA) as novel ways to combine the advantages of both approaches.
Our adjustment methods are based on an auto-regressive AR1 co-kriging model that we integrate with GP priors.
arXiv Detail & Related papers (2023-09-11T12:10:19Z) - Model Predictive Control with Self-supervised Representation Learning [13.225264876433528]
We propose the use of a reconstruction function within the TD-MPC framework, so that the agent can reconstruct the original observation.
Our proposed addition of another loss term leads to improved performance on both state- and image-based tasks.
arXiv Detail & Related papers (2023-04-14T16:02:04Z) - Denoised MDPs: Learning World Models Better Than the World Itself [94.74665254213588]
This work categorizes information out in the wild into four types based on controllability and relation with reward, and formulates useful information as that which is both controllable and reward-relevant.
Experiments on variants of DeepMind Control Suite and RoboDesk demonstrate superior performance of our denoised world model over using raw observations alone.
arXiv Detail & Related papers (2022-06-30T17:59:49Z) - Model Predictive Control for Fluid Human-to-Robot Handovers [50.72520769938633]
Planning motions that take human comfort into account is not a part of the human-robot handover process.
We propose to generate smooth motions via an efficient model-predictive control framework.
We conduct human-to-robot handover experiments on a diverse set of objects with several users.
arXiv Detail & Related papers (2022-03-31T23:08:20Z) - Characterizing and overcoming the greedy nature of learning in
multi-modal deep neural networks [62.48782506095565]
We show that due to the greedy nature of learning in deep neural networks, models tend to rely on just one modality while under-fitting the other modalities.
We propose an algorithm to balance the conditional learning speeds between modalities during training and demonstrate that it indeed addresses the issue of greedy learning.
arXiv Detail & Related papers (2022-02-10T20:11:21Z) - Evaluating model-based planning and planner amortization for continuous
control [79.49319308600228]
We take a hybrid approach, combining model predictive control (MPC) with a learned model and model-free policy learning.
We find that well-tuned model-free agents are strong baselines even for high DoF control problems.
We show that it is possible to distil a model-based planner into a policy that amortizes the planning without any loss of performance.
arXiv Detail & Related papers (2021-10-07T12:00:40Z) - Deep Learning of Koopman Representation for Control [0.0]
The proposed approach relies on the Deep Neural Network based learning of Koopman operator for the purpose of control.
The controller is purely data-driven and does not rely on a priori domain knowledge.
The method is applied to two classic dynamical systems on OpenAI Gym environment to demonstrate the capability.
arXiv Detail & Related papers (2020-10-15T06:41:24Z) - Data Driven Control with Learned Dynamics: Model-Based versus Model-Free
Approach [0.0]
We compare two types of data-driven control methods, representing model-based and model-free approaches.
One is a recently proposed method - Deep Koopman Representation for Control (DKRC), which utilizes a deep neural network to map an unknown nonlinear dynamical system to a high-dimensional linear system.
The other is a classic model-free control method based on an actor-critic architecture - Deep Deterministic Policy Gradient (DDPG), which has been proved to be effective in various dynamical systems.
arXiv Detail & Related papers (2020-06-16T22:18:21Z) - Mining Implicit Entity Preference from User-Item Interaction Data for
Knowledge Graph Completion via Adversarial Learning [82.46332224556257]
We propose a novel adversarial learning approach by leveraging user interaction data for the Knowledge Graph Completion task.
Our generator is isolated from user interaction data, and serves to improve the performance of the discriminator.
To discover implicit entity preference of users, we design an elaborate collaborative learning algorithms based on graph neural networks.
arXiv Detail & Related papers (2020-03-28T05:47:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.