APECS: Adaptive Personalized Control System Architecture
- URL: http://arxiv.org/abs/2503.09624v1
- Date: Mon, 10 Mar 2025 20:11:19 GMT
- Title: APECS: Adaptive Personalized Control System Architecture
- Authors: Marius F. R. Juston, Alex Gisi, William R. Norris, Dustin Nottage, Ahmet Soylemezoglu,
- Abstract summary: This paper presents the Adaptive Personalized Control System (APECS) architecture, a novel framework for human-in-the-loop control.<n>An architecture is developed which defines appropriate constraints for the system objectives.<n>A method for enacting Lipschitz and sector bounds on the resulting controller is derived to ensure desirable control properties.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents the Adaptive Personalized Control System (APECS) architecture, a novel framework for human-in-the-loop control. An architecture is developed which defines appropriate constraints for the system objectives. A method for enacting Lipschitz and sector bounds on the resulting controller is derived to ensure desirable control properties. An analysis of worst-case loss functions and the optimal loss function weighting is made to implement an effective training scheme. Finally, simulations are carried out to demonstrate the effectiveness of the proposed architecture. This architecture resulted in a 4.5% performance increase compared to the human operator and 9% to an unconstrained feedforward neural network trained in the same way.
Related papers
- Neural Network-assisted Interval Reachability for Systems with Control Barrier Function-Based Safe Controllers [1.77513002450736]
Control Barrier Functions (CBFs) have been widely utilized in the design of optimization-based controllers and filters for dynamical systems.
CBF-based controllers offer safety guarantees, but they can compromise the performance of the system.
We propose a computationally efficient interval reachimatability method for performance verification of systems with optimization-based controllers.
arXiv Detail & Related papers (2025-04-11T04:14:55Z) - Unsupervised Parameter Efficient Source-free Post-pretraining [52.27955794126508]
We introduce UpStep, an Unsupervised.<n>Source-free post-pretraining approach to adapt a base model from a source domain to a target domain.<n>We use various general backbone architectures, both supervised and unsupervised, trained on Imagenet as our base model.
arXiv Detail & Related papers (2025-02-28T18:54:51Z) - AI-in-the-Loop Sensing and Communication Joint Design for Edge Intelligence [65.29835430845893]
We propose a framework that enhances edge intelligence through AI-in-the-loop joint sensing and communication.<n>A key contribution of our work is establishing an explicit relationship between validation loss and the system's tunable parameters.<n>We show that our framework reduces communication energy consumption by up to 77 percent and sensing costs measured by the number of samples by up to 52 percent.
arXiv Detail & Related papers (2025-02-14T14:56:58Z) - Structure-Enhanced DRL for Optimal Transmission Scheduling [43.801422320012286]
This paper focuses on the transmission scheduling problem of a remote estimation system.
We develop a structure-enhanced deep reinforcement learning framework for optimal scheduling of the system.
In particular, we propose a structure-enhanced action selection method, which tends to select actions that obey the policy structure.
arXiv Detail & Related papers (2022-12-24T10:18:38Z) - Structure-Enhanced Deep Reinforcement Learning for Optimal Transmission
Scheduling [47.29474858956844]
We develop a structure-enhanced deep reinforcement learning framework for optimal scheduling of a multi-sensor remote estimation system.
In particular, we propose a structure-enhanced action selection method, which tends to select actions that obey the policy structure.
Our numerical results show that the proposed structure-enhanced DRL algorithms can save the training time by 50% and reduce the remote estimation MSE by 10% to 25%.
arXiv Detail & Related papers (2022-11-20T00:13:35Z) - SmoothNets: Optimizing CNN architecture design for differentially
private deep learning [69.10072367807095]
DPSGD requires clipping and noising of per-sample gradients.
This introduces a reduction in model utility compared to non-private training.
We distilled a new model architecture termed SmoothNet, which is characterised by increased robustness to the challenges of DP-SGD training.
arXiv Detail & Related papers (2022-05-09T07:51:54Z) - Optimisation of Structured Neural Controller Based on Continuous-Time
Policy Gradient [2.297079626504224]
This study presents a policy optimisation framework for structured nonlinear control of continuous-time (deterministic) dynamic systems.
The proposed approach prescribes a structure for the controller based on relevant scientific knowledge.
Numerical experiments on aerospace applications illustrate the utility of the structured nonlinear controller optimisation framework.
arXiv Detail & Related papers (2022-01-17T08:06:19Z) - RL-Controller: a reinforcement learning framework for active structural
control [0.0]
We present a novel RL-based approach for designing active controllers by introducing RL-Controller, a flexible and scalable simulation environment.
We show that the proposed framework is easily trainable for a five story benchmark building with 65% reductions on average in inter story drifts.
In a comparative study with LQG active control method, we demonstrate that the proposed model-free algorithm learns more optimal actuator forcing strategies.
arXiv Detail & Related papers (2021-03-13T04:42:13Z) - Reinforcement Learning Control of Robotic Knee with Human in the Loop by
Flexible Policy Iteration [17.365135977882215]
This study fills important voids by introducing innovative features to the policy algorithm.
We show system level performances including convergence of the approximate value function, (sub)optimality of the solution, and stability of the system.
arXiv Detail & Related papers (2020-06-16T09:09:48Z) - Guided Constrained Policy Optimization for Dynamic Quadrupedal Robot
Locomotion [78.46388769788405]
We introduce guided constrained policy optimization (GCPO), an RL framework based upon our implementation of constrained policy optimization (CPPO)
We show that guided constrained RL offers faster convergence close to the desired optimum resulting in an optimal, yet physically feasible, robotic control behavior without the need for precise reward function tuning.
arXiv Detail & Related papers (2020-02-22T10:15:53Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.