Deep networks for system identification: a Survey
- URL: http://arxiv.org/abs/2301.12832v1
- Date: Mon, 30 Jan 2023 12:38:31 GMT
- Title: Deep networks for system identification: a Survey
- Authors: Gianluigi Pillonetto, Aleksandr Aravkin, Daniel Gedon, Lennart Ljung,
Ant\^onio H. Ribeiro, Thomas B. Sch\"on
- Abstract summary: System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
- Score: 56.34005280792013
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning is a topic of considerable current interest. The availability
of massive data collections and powerful software resources has led to an
impressive amount of results in many application areas that reveal essential
but hidden properties of the observations. System identification learns
mathematical descriptions of dynamic systems from input-output data and can
thus benefit from the advances of deep neural networks to enrich the possible
range of models to choose from. For this reason, we provide a survey of deep
learning from a system identification perspective. We cover a wide spectrum of
topics to enable researchers to understand the methods, providing rigorous
practical and theoretical insights into the benefits and challenges of using
them. The main aim of the identified model is to predict new data from previous
observations. This can be achieved with different deep learning based modelling
techniques and we discuss architectures commonly adopted in the literature,
like feedforward, convolutional, and recurrent networks. Their parameters have
to be estimated from past data trying to optimize the prediction performance.
For this purpose, we discuss a specific set of first-order optimization tools
that is emerged as efficient. The survey then draws connections to the
well-studied area of kernel-based methods. They control the data fit by
regularization terms that penalize models not in line with prior assumptions.
We illustrate how to cast them in deep architectures to obtain deep
kernel-based methods. The success of deep learning also resulted in surprising
empirical observations, like the counter-intuitive behaviour of models with
many parameters. We discuss the role of overparameterized models, including
their connection to kernels, as well as implicit regularization mechanisms
which affect generalization, specifically the interesting phenomena of benign
overfitting ...
Related papers
- Advancing Deep Active Learning & Data Subset Selection: Unifying
Principles with Information-Theory Intuitions [3.0539022029583953]
This thesis aims to enhance the practicality of deep learning by improving the label and training efficiency of deep learning models.
We investigate data subset selection techniques, specifically active learning and active sampling, grounded in information-theoretic principles.
arXiv Detail & Related papers (2024-01-09T01:41:36Z) - Visual Prompting Upgrades Neural Network Sparsification: A Data-Model Perspective [64.04617968947697]
We introduce a novel data-model co-design perspective: to promote superior weight sparsity.
Specifically, customized Visual Prompts are mounted to upgrade neural Network sparsification in our proposed VPNs framework.
arXiv Detail & Related papers (2023-12-03T13:50:24Z) - Breaking the Curse of Dimensionality in Deep Neural Networks by Learning
Invariant Representations [1.9580473532948401]
This thesis explores the theoretical foundations of deep learning by studying the relationship between the architecture of these models and the inherent structures found within the data they process.
We ask What drives the efficacy of deep learning algorithms and allows them to beat the so-called curse of dimensionality.
Our methodology takes an empirical approach to deep learning, combining experimental studies with physics-inspired toy models.
arXiv Detail & Related papers (2023-10-24T19:50:41Z) - Rethinking Bayesian Learning for Data Analysis: The Art of Prior and
Inference in Sparsity-Aware Modeling [20.296566563098057]
Sparse modeling for signal processing and machine learning has been at the focus of scientific research for over two decades.
This article reviews some recent advances in incorporating sparsity-promoting priors into three popular data modeling tools.
arXiv Detail & Related papers (2022-05-28T00:43:52Z) - Model-Based Deep Learning: On the Intersection of Deep Learning and
Optimization [101.32332941117271]
Decision making algorithms are used in a multitude of different applications.
Deep learning approaches that use highly parametric architectures tuned from data without relying on mathematical models are becoming increasingly popular.
Model-based optimization and data-centric deep learning are often considered to be distinct disciplines.
arXiv Detail & Related papers (2022-05-05T13:40:08Z) - Leveraging the structure of dynamical systems for data-driven modeling [111.45324708884813]
We consider the impact of the training set and its structure on the quality of the long-term prediction.
We show how an informed design of the training set, based on invariants of the system and the structure of the underlying attractor, significantly improves the resulting models.
arXiv Detail & Related papers (2021-12-15T20:09:20Z) - A neural anisotropic view of underspecification in deep learning [60.119023683371736]
We show that the way neural networks handle the underspecification of problems is highly dependent on the data representation.
Our results highlight that understanding the architectural inductive bias in deep learning is fundamental to address the fairness, robustness, and generalization of these systems.
arXiv Detail & Related papers (2021-04-29T14:31:09Z) - Bayesian graph convolutional neural networks via tempered MCMC [0.41998444721319217]
Deep learning models, such as convolutional neural networks, have long been applied to image and multi-media tasks.
More recently, there has been more attention to unstructured data that can be represented via graphs.
These types of data are often found in health and medicine, social networks, and research data repositories.
arXiv Detail & Related papers (2021-04-17T04:03:25Z) - Model-Based Deep Learning [155.063817656602]
Signal processing, communications, and control have traditionally relied on classical statistical modeling techniques.
Deep neural networks (DNNs) use generic architectures which learn to operate from data, and demonstrate excellent performance.
We are interested in hybrid techniques that combine principled mathematical models with data-driven systems to benefit from the advantages of both approaches.
arXiv Detail & Related papers (2020-12-15T16:29:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.