Information-Theoretic Bounds and Task-Centric Learning Complexity for Real-World Dynamic Nonlinear Systems
- URL: http://arxiv.org/abs/2509.06599v2
- Date: Mon, 22 Sep 2025 13:24:39 GMT
- Title: Information-Theoretic Bounds and Task-Centric Learning Complexity for Real-World Dynamic Nonlinear Systems
- Authors: Sri Satish Krishna Chaitanya Bulusu, Mikko Sillanpää,
- Abstract summary: Dynamic nonlinear systems exhibit distortions arising from coupled static and dynamic effects.<n>This paper presents a theoretical framework grounded in structured decomposition, variance analysis, and task-centric complexity bounds.
- Score: 0.6875312133832079
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dynamic nonlinear systems exhibit distortions arising from coupled static and dynamic effects. Their intertwined nature poses major challenges for data-driven modeling. This paper presents a theoretical framework grounded in structured decomposition, variance analysis, and task-centric complexity bounds. The framework employs a directional lower bound on interactions between measurable system components, extending orthogonality in inner product spaces to structurally asymmetric settings. This bound supports variance inequalities for decomposed systems. Key behavioral indicators are introduced along with a memory finiteness index. A rigorous power-based condition establishes a measurable link between finite memory in realizable systems and the First Law of Thermodynamics. This offers a more foundational perspective than classical bounds based on the Second Law. Building on this foundation, we formulate a `Behavioral Uncertainty Principle,' demonstrating that static and dynamic distortions cannot be minimized simultaneously. We identify that real-world systems seem to resist complete deterministic decomposition due to entangled static and dynamic effects. We also present two general-purpose theorems linking function variance to mean-squared Lipschitz continuity and learning complexity. This yields a model-agnostic, task-aware complexity metric, showing that lower-variance components are inherently easier to learn. These insights explain the empirical benefits of structured residual learning, including improved generalization, reduced parameter count, and lower training cost, as previously observed in power amplifier linearization experiments. The framework is broadly applicable and offers a scalable, theoretically grounded approach to modeling complex dynamic nonlinear systems.
Related papers
- Physics as the Inductive Bias for Causal Discovery [7.9653270330458446]
Causal discovery is often a data-driven paradigm to analyze complex real-world systems.<n>We develop a scalable sparsity-inducing MLE algorithm that exploits causal graph structure for efficient parameter estimation.
arXiv Detail & Related papers (2026-02-03T23:42:01Z) - Random-Matrix-Induced Simplicity Bias in Over-parameterized Variational Quantum Circuits [72.0643009153473]
We show that expressive variational ansatze enter a Haar-like universality class in which both observable expectation values and parameter gradients concentrate exponentially with system size.<n>As a consequence, the hypothesis class induced by such circuits collapses with high probability to a narrow family of near-constant functions.<n>We further show that this collapse is not unavoidable: tensor-structured VQCs, including tensor-network-based and tensor-hypernetwork parameterizations, lie outside the Haar-like universality class.
arXiv Detail & Related papers (2026-01-05T08:04:33Z) - Hierarchical Physics-Embedded Learning for Spatiotemporal Dynamical Systems [12.832325257647128]
We propose a hierarchical physics-embedded learning framework that advances both forwardtemporal prediction and inverse discovery of physical laws.<n>Known physical laws are directly embedded into the models computational graph, guaranteeing physical consistency.<n>By building the framework upon adaptive Neural Operators, we can effectively capture the non-local dependencies and high-order operators characteristic of dynamical systems.
arXiv Detail & Related papers (2025-10-29T09:18:41Z) - Interpretable neural network system identification method for two families of second-order systems based on characteristic curves [0.0]
We propose a unified data-driven framework that combines the mathematical structure of governing differential equations with the flexibility of neural networks (NNs)<n>At the core of our approach is the concept of characteristic curves (CCs), which represent individual nonlinear functions.<n>To demonstrate the versatility of the CC-based formalism, we introduce three identification strategies.
arXiv Detail & Related papers (2025-09-12T18:32:02Z) - Loss-Complexity Landscape and Model Structure Functions [56.01537787608726]
We develop a framework for dualizing the Kolmogorov structure function $h_x(alpha)$.<n>We establish a mathematical analogy between information-theoretic constructs and statistical mechanics.<n>We explicitly prove the Legendre-Fenchel duality between the structure function and free energy.
arXiv Detail & Related papers (2025-07-17T21:31:45Z) - Neural Contraction Metrics with Formal Guarantees for Discrete-Time Nonlinear Dynamical Systems [17.905596843865705]
Contraction metrics provide a powerful framework for analyzing stability, robustness, and convergence of various dynamical systems.<n>However, identifying these metrics for complex nonlinear systems remains an open challenge due to the lack of effective tools.<n>This paper develops verifiable contraction metrics for discrete scalable nonlinear systems.
arXiv Detail & Related papers (2025-04-23T21:27:32Z) - No Equations Needed: Learning System Dynamics Without Relying on Closed-Form ODEs [56.78271181959529]
This paper proposes a conceptual shift to modeling low-dimensional dynamical systems by departing from the traditional two-step modeling process.<n>Instead of first discovering a closed-form equation and then analyzing it, our approach, direct semantic modeling, predicts the semantic representation of the dynamical system.<n>Our approach not only simplifies the modeling pipeline but also enhances the transparency and flexibility of the resulting models.
arXiv Detail & Related papers (2025-01-30T18:36:48Z) - Stability properties of gradient flow dynamics for the symmetric low-rank matrix factorization problem [22.648448759446907]
We show that a low-rank factorization serves as a building block in many learning tasks.
We offer new insight into the shape of the trajectories associated with local search parts of the dynamics.
arXiv Detail & Related papers (2024-11-24T20:05:10Z) - Learning Controlled Stochastic Differential Equations [61.82896036131116]
This work proposes a novel method for estimating both drift and diffusion coefficients of continuous, multidimensional, nonlinear controlled differential equations with non-uniform diffusion.
We provide strong theoretical guarantees, including finite-sample bounds for (L2), (Linfty), and risk metrics, with learning rates adaptive to coefficients' regularity.
Our method is available as an open-source Python library.
arXiv Detail & Related papers (2024-11-04T11:09:58Z) - Towards Understanding Generalization via Decomposing Excess Risk
Dynamics [13.4379473119565]
We analyze the generalization dynamics to derive algorithm-dependent bounds, e.g., stability.
Inspired by the observation that neural networks show a slow convergence rate when fitting noise, we propose decomposing the excess risk dynamics.
Under the decomposition framework, the new bound accords better with the theoretical and empirical evidence compared to the stability-based bound and uniform convergence bound.
arXiv Detail & Related papers (2021-06-11T03:42:45Z) - Physics-informed Spline Learning for Nonlinear Dynamics Discovery [8.546520029145853]
We propose a Physics-informed Spline Learning framework to discover parsimonious governing equations for nonlinear dynamics.
The framework is based on sparsely sampled noisy data.
The efficacy and superiority of the proposed method has been demonstrated by multiple well-known nonlinear dynamical systems.
arXiv Detail & Related papers (2021-05-05T23:32:43Z) - Learning Theory for Inferring Interaction Kernels in Second-Order
Interacting Agent Systems [17.623937769189364]
We develop a complete learning theory which establishes strong consistency and optimal nonparametric min-max rates of convergence for the estimators.
The numerical algorithm presented to build the estimators is parallelizable, performs well on high-dimensional problems, and is demonstrated on complex dynamical systems.
arXiv Detail & Related papers (2020-10-08T02:07:53Z) - Euclideanizing Flows: Diffeomorphic Reduction for Learning Stable
Dynamical Systems [74.80320120264459]
We present an approach to learn such motions from a limited number of human demonstrations.
The complex motions are encoded as rollouts of a stable dynamical system.
The efficacy of this approach is demonstrated through validation on an established benchmark as well demonstrations collected on a real-world robotic system.
arXiv Detail & Related papers (2020-05-27T03:51:57Z) - On dissipative symplectic integration with applications to
gradient-based optimization [77.34726150561087]
We propose a geometric framework in which discretizations can be realized systematically.
We show that a generalization of symplectic to nonconservative and in particular dissipative Hamiltonian systems is able to preserve rates of convergence up to a controlled error.
arXiv Detail & Related papers (2020-04-15T00:36:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.