Inductive biases and Self Supervised Learning in modelling a physical
heating system
- URL: http://arxiv.org/abs/2104.11478v1
- Date: Fri, 23 Apr 2021 08:50:41 GMT
- Title: Inductive biases and Self Supervised Learning in modelling a physical
heating system
- Authors: Cristian Vicas
- Abstract summary: In this paper I infer inductive biases about a physical system.
I use these biases to derive a new neural network architecture that can model this real system.
The proposed architecture family called Delay can be used in a real scenario to control systems with delayed responses.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Model Predictive Controllers (MPC) require a good model for the controlled
process. In this paper I infer inductive biases about a physical system. I use
these biases to derive a new neural network architecture that can model this
real system that has noise and inertia. The main inductive biases exploited
here are: the delayed impact of some inputs on the system and the separability
between the temporal component and how the inputs interact to produce the
output of a system. The inputs are independently delayed using shifted
convolutional kernels. Feature interactions are modelled using a fully
connected network that does not have access to temporal information. The
available data and the problem setup allow the usage of Self Supervised
Learning in order to train the models. The baseline architecture is an
Attention based Reccurent network adapted to work with MPC like inputs. The
proposed networks are faster, better at exploiting larger data volumes and are
almost as good as baseline networks in terms of prediction performance. The
proposed architecture family called Delay can be used in a real scenario to
control systems with delayed responses with respect to its controls or inputs.
Ablation studies show that the presence of delay kernels are vital to obtain
any learning in proposed architecture. Code and some experimental data are
available online.
Related papers
- Domain-decoupled Physics-informed Neural Networks with Closed-form Gradients for Fast Model Learning of Dynamical Systems [2.8730926763860687]
Physics-informed neural networks (PINNs) are trained using physical equations and can incorporate unmodeled effects by learning from data.
We introduce the domain-decoupled physics-informed neural network (DD-PINN) to address current limitations of PINC in handling large and complex nonlinear dynamical systems.
arXiv Detail & Related papers (2024-08-27T10:54:51Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - An FPGA Architecture for Online Learning using the Tsetlin Machine [5.140342614848069]
This paper proposes a novel field-programmable gate-array infrastructure for online learning.
It implements a low-complexity machine learning algorithm called the Tsetlin Machine.
We present use cases for online learning using the proposed infrastructure and demonstrate the energy/performance/accuracy trade-offs.
arXiv Detail & Related papers (2023-06-01T13:33:26Z) - Brain-Inspired Spiking Neural Network for Online Unsupervised Time
Series Prediction [13.521272923545409]
We present a novel Continuous Learning-based Unsupervised Recurrent Spiking Neural Network Model (CLURSNN)
CLURSNN makes online predictions by reconstructing the underlying dynamical system using Random Delay Embedding.
We show that the proposed online time series prediction methodology outperforms state-of-the-art DNN models when predicting an evolving Lorenz63 dynamical system.
arXiv Detail & Related papers (2023-04-10T16:18:37Z) - Learning Flow Functions from Data with Applications to Nonlinear
Oscillators [0.0]
We show that learning the flow function is equivalent to learning the input-to-state map of a discrete-time dynamical system.
This motivates the use of an RNN together with encoder and decoder networks which map the state of the system to the hidden state of the RNN and back.
arXiv Detail & Related papers (2023-03-29T13:04:04Z) - Deep networks for system identification: a Survey [56.34005280792013]
System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
arXiv Detail & Related papers (2023-01-30T12:38:31Z) - Leveraging the structure of dynamical systems for data-driven modeling [111.45324708884813]
We consider the impact of the training set and its structure on the quality of the long-term prediction.
We show how an informed design of the training set, based on invariants of the system and the structure of the underlying attractor, significantly improves the resulting models.
arXiv Detail & Related papers (2021-12-15T20:09:20Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Model-Based Deep Learning [155.063817656602]
Signal processing, communications, and control have traditionally relied on classical statistical modeling techniques.
Deep neural networks (DNNs) use generic architectures which learn to operate from data, and demonstrate excellent performance.
We are interested in hybrid techniques that combine principled mathematical models with data-driven systems to benefit from the advantages of both approaches.
arXiv Detail & Related papers (2020-12-15T16:29:49Z) - Deep Learning for Ultra-Reliable and Low-Latency Communications in 6G
Networks [84.2155885234293]
We first summarize how to apply data-driven supervised deep learning and deep reinforcement learning in URLLC.
To address these open problems, we develop a multi-level architecture that enables device intelligence, edge intelligence, and cloud intelligence for URLLC.
arXiv Detail & Related papers (2020-02-22T14:38:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.