Thermodynamic Machine Learning through Maximum Work Production
- URL: http://arxiv.org/abs/2006.15416v3
- Date: Mon, 12 Apr 2021 17:04:24 GMT
- Title: Thermodynamic Machine Learning through Maximum Work Production
- Authors: A. B. Boyd, J. P. Crutchfield, and M. Gu
- Abstract summary: We introduce the thermodynamic principle that work production is the most relevant performance metric for an adaptive physical agent.
We show that selecting the maximum-work agent for given environmental data corresponds to finding the maximum-likelihood model.
In this way, work emerges as an organizing principle that underlies learning in adaptive thermodynamic systems.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Adaptive systems -- such as a biological organism gaining survival advantage,
an autonomous robot executing a functional task, or a motor protein
transporting intracellular nutrients -- must model the regularities and
stochasticity in their environments to take full advantage of thermodynamic
resources. Analogously, but in a purely computational realm, machine learning
algorithms estimate models to capture predictable structure and identify
irrelevant noise in training data. This happens through optimization of
performance metrics, such as model likelihood. If physically implemented, is
there a sense in which computational models estimated through machine learning
are physically preferred? We introduce the thermodynamic principle that work
production is the most relevant performance metric for an adaptive physical
agent and compare the results to the maximum-likelihood principle that guides
machine learning. Within the class of physical agents that most efficiently
harvest energy from their environment, we demonstrate that an efficient agent's
model explicitly determines its architecture and how much useful work it
harvests from the environment. We then show that selecting the maximum-work
agent for given environmental data corresponds to finding the
maximum-likelihood model. This establishes an equivalence between
nonequilibrium thermodynamics and dynamic learning. In this way, work
maximization emerges as an organizing principle that underlies learning in
adaptive thermodynamic systems.
Related papers
- The Work Capacity of Channels with Memory: Maximum Extractable Work in Percept-Action Loops [0.5999777817331317]
We analyze the thermodynamics of information processing in percept-action loops.
We introduce the concept of work capacity -- the maximum rate at which an agent can expect to extract work from its environment.
arXiv Detail & Related papers (2025-04-08T16:54:20Z) - Understanding Machine Learning Paradigms through the Lens of Statistical Thermodynamics: A tutorial [0.0]
The tutorial delves into advanced techniques like entropy, free energy, and variational inference which are utilized in machine learning.
We show how an in-depth comprehension of physical systems' behavior can yield more effective and dependable machine learning models.
arXiv Detail & Related papers (2024-11-24T18:20:05Z) - DiffuseBot: Breeding Soft Robots With Physics-Augmented Generative
Diffusion Models [102.13968267347553]
We present DiffuseBot, a physics-augmented diffusion model that generates soft robot morphologies capable of excelling in a wide spectrum of tasks.
We showcase a range of simulated and fabricated robots along with their capabilities.
arXiv Detail & Related papers (2023-11-28T18:58:48Z) - STORM: Efficient Stochastic Transformer based World Models for
Reinforcement Learning [82.03481509373037]
Recently, model-based reinforcement learning algorithms have demonstrated remarkable efficacy in visual input environments.
We introduce Transformer-based wORld Model (STORM), an efficient world model architecture that combines strong modeling and generation capabilities.
Storm achieves a mean human performance of $126.7%$ on the Atari $100$k benchmark, setting a new record among state-of-the-art methods.
arXiv Detail & Related papers (2023-10-14T16:42:02Z) - A Comparative Study of Machine Learning Algorithms for Anomaly Detection
in Industrial Environments: Performance and Environmental Impact [62.997667081978825]
This study seeks to address the demands of high-performance machine learning models with environmental sustainability.
Traditional machine learning algorithms, such as Decision Trees and Random Forests, demonstrate robust efficiency and performance.
However, superior outcomes were obtained with optimised configurations, albeit with a commensurate increase in resource consumption.
arXiv Detail & Related papers (2023-07-01T15:18:00Z) - Dynamic-Resolution Model Learning for Object Pile Manipulation [33.05246884209322]
We investigate how to learn dynamic and adaptive representations at different levels of abstraction to achieve the optimal trade-off between efficiency and effectiveness.
Specifically, we construct dynamic-resolution particle representations of the environment and learn a unified dynamics model using graph neural networks (GNNs)
We show that our method achieves significantly better performance than state-of-the-art fixed-resolution baselines at the gathering, sorting, and redistribution of granular object piles.
arXiv Detail & Related papers (2023-06-29T05:51:44Z) - Pessimism meets VCG: Learning Dynamic Mechanism Design via Offline
Reinforcement Learning [114.36124979578896]
We design a dynamic mechanism using offline reinforcement learning algorithms.
Our algorithm is based on the pessimism principle and only requires a mild assumption on the coverage of the offline data set.
arXiv Detail & Related papers (2022-05-05T05:44:26Z) - Gradient-Based Trajectory Optimization With Learned Dynamics [80.41791191022139]
We use machine learning techniques to learn a differentiable dynamics model of the system from data.
We show that a neural network can model highly nonlinear behaviors accurately for large time horizons.
In our hardware experiments, we demonstrate that our learned model can represent complex dynamics for both the Spot and Radio-controlled (RC) car.
arXiv Detail & Related papers (2022-04-09T22:07:34Z) - Model Predictive Actor-Critic: Accelerating Robot Skill Acquisition with
Deep Reinforcement Learning [42.525696463089794]
Model Predictive Actor-Critic (MoPAC) is a hybrid model-based/model-free method that combines model predictive rollouts with policy optimization as to mitigate model bias.
MoPAC guarantees optimal skill learning up to an approximation error and reduces necessary physical interaction with the environment.
arXiv Detail & Related papers (2021-03-25T13:50:24Z) - Tensor network approaches for learning non-linear dynamical laws [0.0]
We show that various physical constraints can be captured via tensor network based parameterizations for the governing equation.
We provide a physics-informed approach to recovering structured dynamical laws from data, which adaptively balances the need for expressivity and scalability.
arXiv Detail & Related papers (2020-02-27T19:02:40Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.