Structured model selection via $\ell_1-\ell_2$ optimization
- URL: http://arxiv.org/abs/2305.17467v2
- Date: Tue, 30 May 2023 00:54:10 GMT
- Title: Structured model selection via $\ell_1-\ell_2$ optimization
- Authors: Xiaofan Lu, Linan Zhang and Hongjin He
- Abstract summary: We develop a learning approach for identifying structured dynamical systems.
We show that if the set of candidate functions forms a bounded system, the recovery is stable and is bounded.
- Score: 1.933681537640272
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automated model selection is an important application in science and
engineering. In this work, we develop a learning approach for identifying
structured dynamical systems from undersampled and noisy spatiotemporal data.
The learning is performed by a sparse least-squares fitting over a large set of
candidate functions via a nonconvex $\ell_1-\ell_2$ sparse optimization solved
by the alternating direction method of multipliers. Using a Bernstein-like
inequality with a coherence condition, we show that if the set of candidate
functions forms a structured random sampling matrix of a bounded orthogonal
system, the recovery is stable and the error is bounded. The learning approach
is validated on synthetic data generated by the viscous Burgers' equation and
two reaction-diffusion equations. The computational results demonstrate the
theoretical guarantees of success and the efficiency with respect to the
ambient dimension and the number of candidate functions.
Related papers
- Accelerated zero-order SGD under high-order smoothness and overparameterized regime [79.85163929026146]
We present a novel gradient-free algorithm to solve convex optimization problems.
Such problems are encountered in medicine, physics, and machine learning.
We provide convergence guarantees for the proposed algorithm under both types of noise.
arXiv Detail & Related papers (2024-11-21T10:26:17Z) - Bisimulation Learning [55.859538562698496]
We compute finite bisimulations of state transition systems with large, possibly infinite state space.
Our technique yields faster verification results than alternative state-of-the-art tools in practice.
arXiv Detail & Related papers (2024-05-24T17:11:27Z) - Comparison of Single- and Multi- Objective Optimization Quality for
Evolutionary Equation Discovery [77.34726150561087]
Evolutionary differential equation discovery proved to be a tool to obtain equations with less a priori assumptions.
The proposed comparison approach is shown on classical model examples -- Burgers equation, wave equation, and Korteweg - de Vries equation.
arXiv Detail & Related papers (2023-06-29T15:37:19Z) - Learning Unnormalized Statistical Models via Compositional Optimization [73.30514599338407]
Noise-contrastive estimation(NCE) has been proposed by formulating the objective as the logistic loss of the real data and the artificial noise.
In this paper, we study it a direct approach for optimizing the negative log-likelihood of unnormalized models.
arXiv Detail & Related papers (2023-06-13T01:18:16Z) - Score-based Diffusion Models in Function Space [140.792362459734]
Diffusion models have recently emerged as a powerful framework for generative modeling.
We introduce a mathematically rigorous framework called Denoising Diffusion Operators (DDOs) for training diffusion models in function space.
We show that the corresponding discretized algorithm generates accurate samples at a fixed cost independent of the data resolution.
arXiv Detail & Related papers (2023-02-14T23:50:53Z) - WeakIdent: Weak formulation for Identifying Differential Equations using
Narrow-fit and Trimming [5.027714423258538]
We propose a general and robust framework to recover differential equations using a weak formulation.
For each sparsity level, Subspace Pursuit is utilized to find an initial set of support from the large dictionary.
The proposed method gives a robust recovery of the coefficients, and a significant denoising effect which can handle up to $100%$ noise-to-signal ratio.
arXiv Detail & Related papers (2022-11-06T14:33:22Z) - Sparse PCA via $l_{2,p}$-Norm Regularization for Unsupervised Feature
Selection [138.97647716793333]
We propose a simple and efficient unsupervised feature selection method, by combining reconstruction error with $l_2,p$-norm regularization.
We present an efficient optimization algorithm to solve the proposed unsupervised model, and analyse the convergence and computational complexity of the algorithm theoretically.
arXiv Detail & Related papers (2020-12-29T04:08:38Z) - DiffPrune: Neural Network Pruning with Deterministic Approximate Binary
Gates and $L_0$ Regularization [0.0]
Modern neural network architectures typically have many millions of parameters and can be pruned significantly without substantial loss in effectiveness.
The contribution of this work is two-fold.
The first is a method for approximating a multivariate Bernoulli random variable by means of a deterministic and differentiable transformation of any real-valued random variable.
The second is a method for model selection by element-wise parameters with approximate binary gates that may be computed deterministically or multiplicationally and take on exact zero values.
arXiv Detail & Related papers (2020-12-07T13:08:56Z) - A non-autonomous equation discovery method for time signal
classification [1.933681537640272]
We develop a framework for analyzing time signals based on non-autonomous dynamical equations.
We show how gradients can be efficiently computed using the adjoint method.
We also demonstrate how the proposed method yields interpretability in the form of phase portraits.
arXiv Detail & Related papers (2020-11-22T20:03:46Z) - Semi-analytic approximate stability selection for correlated data in
generalized linear models [3.42658286826597]
We propose a novel approximate inference algorithm that can conduct Stability Selection without the repeated fitting.
The algorithm is based on the replica method of statistical mechanics and vector approximate message passing of information theory.
Numerical experiments indicate that the algorithm exhibits fast convergence and high approximation accuracy for both synthetic and real-world data.
arXiv Detail & Related papers (2020-03-19T10:43:12Z) - Analysis of Bayesian Inference Algorithms by the Dynamical Functional
Approach [2.8021833233819486]
We analyze an algorithm for approximate inference with large Gaussian latent variable models in a student-trivial scenario.
For the case of perfect data-model matching, the knowledge of static order parameters derived from the replica method allows us to obtain efficient algorithmic updates.
arXiv Detail & Related papers (2020-01-14T17:22:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.