The Origin and Evolution of Information Handling
- URL: http://arxiv.org/abs/2404.04374v5
- Date: Fri, 07 Feb 2025 16:54:53 GMT
- Title: The Origin and Evolution of Information Handling
- Authors: Amahury Jafet López-Díaz, Hiroki Sayama, Carlos Gershenson,
- Abstract summary: Information-first approach integrates Hofmeyr's (F, A)-systems with temporal parametrization and multiscale causality.
Our model traces the evolution of information handling from simple reaction networks that recognize regular languages to self-replicating chemical systems with memory and anticipatory capabilities.
- Score: 0.6963971634605796
- License:
- Abstract: Understanding the emergence and evolution of information handling is essential for unraveling the origins of life. Traditional genetic-first and metabolism-first models fall short in explaining how instructional information control systems naturally arise from molecular dynamics. To address this gap, we adopt an information-first approach, integrating Hofmeyr's (F, A)-systems -- an extension of Rosen's (M, R)-systems -- with temporal parametrization and multiscale causality. These models, which embody closure to efficient causation while remaining open to formal causation, provide a robust framework for primitive autopoiesis, anticipation, and adaptation. We establish a formal equivalence between extended (F, A)-systems and communicating X-machines, resolving self-referential challenges and demonstrating the hypercomputational nature of life processes. Our stepwise model traces the evolution of information handling from simple reaction networks that recognize regular languages to self-replicating chemical systems with memory and anticipatory capabilities. This transition from analog to digital architectures enhances evolutionary robustness and aligns with experimental evidence suggesting that chemical computation does not require life-specific chemistry. Furthermore, we incorporate open-ended evolutionary dynamics driven by computational undecidability and irreducibility, reinforcing the necessity of unconventional computing frameworks. This computational enactivist perspective provides a cohesive theoretical basis for a recently proposed trialectic between autopoiesis, anticipation and adaptation in order to solve the problem of relevance. By highlighting the critical role of hypercomputational processes in life's emergence and evolution, our framework offers new insights into the fundamental principles underlying biological information processing.
Related papers
- GENERator: A Long-Context Generative Genomic Foundation Model [66.46537421135996]
We present a generative genomic foundation model featuring a context length of 98k base pairs (bp) and 1.2B parameters.
The model adheres to the central dogma of molecular biology, accurately generating protein-coding sequences.
It also shows significant promise in sequence optimization, particularly through the prompt-responsive generation of promoter sequences.
arXiv Detail & Related papers (2025-02-11T05:39:49Z) - No Foundations without Foundations -- Why semi-mechanistic models are essential for regulatory biology [5.925258390690544]
We argue that genuine "foundation models" of regulatory biology will remain out of reach unless guided by frameworks that integrate mechanistic insight with principled experimental design.
We present one such ground-up, semi-mechanistic framework that unifies perturbation-based experimental designs.
arXiv Detail & Related papers (2025-01-31T14:43:16Z) - No Equations Needed: Learning System Dynamics Without Relying on Closed-Form ODEs [56.78271181959529]
This paper proposes a conceptual shift to modeling low-dimensional dynamical systems by departing from the traditional two-step modeling process.
Instead of first discovering a closed-form equation and then analyzing it, our approach, direct semantic modeling, predicts the semantic representation of the dynamical system.
Our approach not only simplifies the modeling pipeline but also enhances the transparency and flexibility of the resulting models.
arXiv Detail & Related papers (2025-01-30T18:36:48Z) - Rethinking Cognition: Morphological Info-Computation and the Embodied Paradigm in Life and Artificial Intelligence [1.14219428942199]
This study aims to place Lorenzo Magnanis Eco-Cognitive Computationalism within the broader context of current work on information, computation, and cognition.
We model cognition as a web of concurrent morphological computations, driven by processes of self-assembly, self-organisation, and autopoiesis across physical, chemical, and biological domains.
arXiv Detail & Related papers (2024-12-01T10:04:53Z) - Deep Signature: Characterization of Large-Scale Molecular Dynamics [29.67824486345836]
Deep Signature is a novel computationally tractable framework that characterizes complex dynamics and interatomic interactions.
Our approach incorporates soft spectral clustering that locally aggregates cooperative dynamics to reduce the size of the system, as well as signature transform to provide a global characterization of the non-smooth interactive dynamics.
arXiv Detail & Related papers (2024-10-03T16:37:48Z) - Causal machine learning for single-cell genomics [94.28105176231739]
We discuss the application of machine learning techniques to single-cell genomics and their challenges.
We first present the model that underlies most of current causal approaches to single-cell biology.
We then identify open problems in the application of causal approaches to single-cell data.
arXiv Detail & Related papers (2023-10-23T13:35:24Z) - A Neuro-mimetic Realization of the Common Model of Cognition via Hebbian
Learning and Free Energy Minimization [55.11642177631929]
Large neural generative models are capable of synthesizing semantically rich passages of text or producing complex images.
We discuss the COGnitive Neural GENerative system, such an architecture that casts the Common Model of Cognition.
arXiv Detail & Related papers (2023-10-14T23:28:48Z) - Towards Predicting Equilibrium Distributions for Molecular Systems with
Deep Learning [60.02391969049972]
We introduce a novel deep learning framework, called Distributional Graphormer (DiG), in an attempt to predict the equilibrium distribution of molecular systems.
DiG employs deep neural networks to transform a simple distribution towards the equilibrium distribution, conditioned on a descriptor of a molecular system.
arXiv Detail & Related papers (2023-06-08T17:12:08Z) - Evolution TANN and the discovery of the internal variables and evolution
equations in solid mechanics [0.0]
We propose a new approach which allows, for the first time, to decouple the material representation from the incremental formulation.
Inspired by the Thermodynamics-based Artificial Neural Networks (TANN) and the theory of the internal variables, the evolution TANN (eTANN) are continuous-time.
Key feature of the proposed approach is the discovery of the evolution equations of the internal variables in the form of ordinary differential equations.
arXiv Detail & Related papers (2022-09-27T09:25:55Z) - Discovering Latent Causal Variables via Mechanism Sparsity: A New
Principle for Nonlinear ICA [81.4991350761909]
Independent component analysis (ICA) refers to an ensemble of methods which formalize this goal and provide estimation procedure for practical application.
We show that the latent variables can be recovered up to a permutation if one regularizes the latent mechanisms to be sparse.
arXiv Detail & Related papers (2021-07-21T14:22:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.