Control-Theoretic Techniques for Online Adaptation of Deep Neural
Networks in Dynamical Systems
- URL: http://arxiv.org/abs/2402.00761v1
- Date: Thu, 1 Feb 2024 16:51:11 GMT
- Title: Control-Theoretic Techniques for Online Adaptation of Deep Neural
Networks in Dynamical Systems
- Authors: Jacob G. Elkins and Farbod Fahimi
- Abstract summary: Deep neural networks (DNNs) are currently the primary tool in modern artificial intelligence, machine learning, and data science.
In many applications, DNNs are trained offline, through supervised learning or reinforcement learning, and deployed online for inference.
We propose using techniques from control theory to update DNN parameters online.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks (DNNs), trained with gradient-based optimization and
backpropagation, are currently the primary tool in modern artificial
intelligence, machine learning, and data science. In many applications, DNNs
are trained offline, through supervised learning or reinforcement learning, and
deployed online for inference. However, training DNNs with standard
backpropagation and gradient-based optimization gives no intrinsic performance
guarantees or bounds on the DNN, which is essential for applications such as
controls. Additionally, many offline-training and online-inference problems,
such as sim2real transfer of reinforcement learning policies, experience domain
shift from the training distribution to the real-world distribution. To address
these stability and transfer learning issues, we propose using techniques from
control theory to update DNN parameters online. We formulate the
fully-connected feedforward DNN as a continuous-time dynamical system, and we
propose novel last-layer update laws that guarantee desirable error convergence
under various conditions on the time derivative of the DNN input vector. We
further show that training the DNN under spectral normalization controls the
upper bound of the error trajectories of the online DNN predictions, which is
desirable when numerically differentiated quantities or noisy state
measurements are input to the DNN. The proposed online DNN adaptation laws are
validated in simulation to learn the dynamics of the Van der Pol system under
domain shift, where parameters are varied in inference from the training
dataset. The simulations demonstrate the effectiveness of using
control-theoretic techniques to derive performance improvements and guarantees
in DNN-based learning systems.
Related papers
- Gradient Transformation: Towards Efficient and Model-Agnostic Unlearning for Dynamic Graph Neural Networks [66.70786325911124]
Graph unlearning has emerged as an essential tool for safeguarding user privacy and mitigating the negative impacts of undesirable data.
With the increasing prevalence of DGNNs, it becomes imperative to investigate the implementation of dynamic graph unlearning.
We propose an effective, efficient, model-agnostic, and post-processing method to implement DGNN unlearning.
arXiv Detail & Related papers (2024-05-23T10:26:18Z) - Enhancing Deep Neural Network Training Efficiency and Performance through Linear Prediction [0.0]
Deep neural networks (DNN) have achieved remarkable success in various fields, including computer vision and natural language processing.
This paper aims to propose a method to optimize the training effectiveness of DNN, with the goal of improving model performance.
arXiv Detail & Related papers (2023-10-17T03:11:30Z) - Brain-Inspired Spiking Neural Network for Online Unsupervised Time
Series Prediction [13.521272923545409]
We present a novel Continuous Learning-based Unsupervised Recurrent Spiking Neural Network Model (CLURSNN)
CLURSNN makes online predictions by reconstructing the underlying dynamical system using Random Delay Embedding.
We show that the proposed online time series prediction methodology outperforms state-of-the-art DNN models when predicting an evolving Lorenz63 dynamical system.
arXiv Detail & Related papers (2023-04-10T16:18:37Z) - Adversarial Learning Networks: Source-free Unsupervised Domain
Incremental Learning [0.0]
In a non-stationary environment, updating a DNN model requires parameter re-training or model fine-tuning.
We propose an unsupervised source-free method to update DNN classification models.
Unlike existing methods, our approach can update a DNN model incrementally for non-stationary source and target tasks without storing past training data.
arXiv Detail & Related papers (2023-01-28T02:16:13Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Online Training Through Time for Spiking Neural Networks [66.7744060103562]
Spiking neural networks (SNNs) are promising brain-inspired energy-efficient models.
Recent progress in training methods has enabled successful deep SNNs on large-scale tasks with low latency.
We propose online training through time (OTTT) for SNNs, which is derived from BPTT to enable forward-in-time learning.
arXiv Detail & Related papers (2022-10-09T07:47:56Z) - Deep neural network based adaptive learning for switched systems [0.3222802562733786]
We present a deep neural network based adaptive learning (DNN-AL) approach for switched systems.
observed datasets are adaptively decomposed into subsets, such as no structural changes within each subset.
Network parameters at previous iteration steps are reused to initialize networks for the later iteration steps.
arXiv Detail & Related papers (2022-07-11T04:51:58Z) - slimTrain -- A Stochastic Approximation Method for Training Separable
Deep Neural Networks [2.4373900721120285]
DeepTrain networks (DNNs) have shown their success as high-dimensional neural function approximators in many applications.
We propose slimTrain, a modest optimization method for training DNNs with reduced sensitivity to the choice hyper-dimensional datasets.
arXiv Detail & Related papers (2021-09-28T19:31:57Z) - Wide and Deep Graph Neural Network with Distributed Online Learning [174.8221510182559]
Graph neural networks (GNNs) are naturally distributed architectures for learning representations from network data.
Online learning can be leveraged to retrain GNNs at testing time to overcome this issue.
This paper develops the Wide and Deep GNN (WD-GNN), a novel architecture that can be updated with distributed online learning mechanisms.
arXiv Detail & Related papers (2021-07-19T23:56:48Z) - A Meta-Learning Approach to the Optimal Power Flow Problem Under
Topology Reconfigurations [69.73803123972297]
We propose a DNN-based OPF predictor that is trained using a meta-learning (MTL) approach.
The developed OPF-predictor is validated through simulations using benchmark IEEE bus systems.
arXiv Detail & Related papers (2020-12-21T17:39:51Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.