Collective advantages in finite-time thermodynamics
- URL: http://arxiv.org/abs/2306.16534v3
- Date: Fri, 20 Oct 2023 14:54:21 GMT
- Title: Collective advantages in finite-time thermodynamics
- Authors: Alberto Rolandi, Paolo Abiuso, Mart\'i Perarnau-Llobet
- Abstract summary: We show that $W_rm disspropto Nx$ can be dramatically reduced by considering collective protocols in which interactions are suitably created along the protocol.
As an application of these results, we focus on the erasure of information in finite time and prove a faster convergence to Landauer's bound.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A central task in finite-time thermodynamics is to minimize the excess or
dissipated work $W_{\rm diss}$ when manipulating the state of a system immersed
in a thermal bath. We consider this task for an $N$-body system whose
constituents are identical and uncorrelated at the beginning and end of the
process. In the regime of slow but finite-time processes, we show that $W_{\rm
diss}$ can be dramatically reduced by considering collective protocols in which
interactions are suitably created along the protocol. This can even lead to a
sub-linear growth of $W_{\rm diss}$ with $N$: $W_{\rm diss}\propto N^x$ with
$x<1$; to be contrasted to the expected $W_{\rm diss}\propto N$ satisfied in
any non-interacting protocol. We derive the fundamental limits to such
collective advantages and show that $x=0$ is in principle possible, however it
requires long-range interactions. We explore collective processes with spin
models featuring two-body interactions and achieve noticeable gains under
realistic levels of control in simple interaction architectures. As an
application of these results, we focus on the erasure of information in finite
time and prove a faster convergence to Landauer's bound.
Related papers
- Near-Optimal Online Learning for Multi-Agent Submodular Coordination: Tight Approximation and Communication Efficiency [52.60557300927007]
We present a $textbfMA-OSMA$ algorithm to transfer the discrete submodular problem into a continuous optimization.
We also introduce a projection-free $textbfMA-OSEA$ algorithm, which effectively utilizes the KL divergence by mixing a uniform distribution.
Our algorithms significantly improve the $(frac11+c)$-approximation provided by the state-of-the-art OSG algorithm.
arXiv Detail & Related papers (2025-02-07T15:57:56Z) - Minimizing Dissipation via Interacting Environments: Quadratic Convergence to Landauer Bound [0.0]
We prove that for any non-interacting $n$-particle reservoir, the entropy production $Sigma$ decays at most linearly with $n$.
We derive a cooling protocol in which $Sigma propto 1/n2$, which is in fact the best possible scaling.
arXiv Detail & Related papers (2024-11-01T18:00:08Z) - Projection by Convolution: Optimal Sample Complexity for Reinforcement Learning in Continuous-Space MDPs [56.237917407785545]
We consider the problem of learning an $varepsilon$-optimal policy in a general class of continuous-space Markov decision processes (MDPs) having smooth Bellman operators.
Key to our solution is a novel projection technique based on ideas from harmonic analysis.
Our result bridges the gap between two popular but conflicting perspectives on continuous-space MDPs.
arXiv Detail & Related papers (2024-05-10T09:58:47Z) - Extraction of Work via a Thermalization Protocol [0.0]
We show that it is possible to exploit a thermalization process to extract work from a resource system $R$ to a bipartite system $S$.
We find the theoretical bounds of the protocol in the general case and show that when applied to the Rabi model it gives rise to a satisfactory extraction of work and efficiency.
arXiv Detail & Related papers (2023-09-08T08:01:40Z) - Settling the Sample Complexity of Online Reinforcement Learning [92.02082223856479]
We show how to achieve minimax-optimal regret without incurring any burn-in cost.
We extend our theory to unveil the influences of problem-dependent quantities like the optimal value/cost and certain variances.
arXiv Detail & Related papers (2023-07-25T15:42:11Z) - Cooperative Multi-Agent Reinforcement Learning: Asynchronous
Communication and Linear Function Approximation [77.09836892653176]
We study multi-agent reinforcement learning in the setting of episodic Markov decision processes.
We propose a provably efficient algorithm based on value that enable asynchronous communication.
We show that a minimal $Omega(dM)$ communication complexity is required to improve the performance through collaboration.
arXiv Detail & Related papers (2023-05-10T20:29:29Z) - Reward-Mixing MDPs with a Few Latent Contexts are Learnable [75.17357040707347]
We consider episodic reinforcement learning in reward-mixing Markov decision processes (RMMDPs)
Our goal is to learn a near-optimal policy that nearly maximizes the $H$ time-step cumulative rewards in such a model.
arXiv Detail & Related papers (2022-10-05T22:52:00Z) - DADAO: Decoupled Accelerated Decentralized Asynchronous Optimization [0.0]
DADAO is the first decentralized, accelerated, asynchronous, primal, first-order algorithm to minimize a sum of $L$-smooth and $mu$-strongly convex functions distributed over a given network of size $n$.
We show that our algorithm requires $mathcalO(nsqrtchisqrtfracLmulog(frac1epsilon)$ local and only $mathcalO(nsqrtchisqrtfracLmulog(
arXiv Detail & Related papers (2022-07-26T08:47:54Z) - Provably Efficient Offline Reinforcement Learning with Trajectory-Wise
Reward [66.81579829897392]
We propose a novel offline reinforcement learning algorithm called Pessimistic vAlue iteRaTion with rEward Decomposition (PARTED)
PARTED decomposes the trajectory return into per-step proxy rewards via least-squares-based reward redistribution, and then performs pessimistic value based on the learned proxy reward.
To the best of our knowledge, PARTED is the first offline RL algorithm that is provably efficient in general MDP with trajectory-wise reward.
arXiv Detail & Related papers (2022-06-13T19:11:22Z) - Finite-time bounds on the probabilistic violation of the second law of
thermodynamics [0.0]
We show that finite-time protocols converge to Jarzynski's bound at a rate slower than $1/sqrttau$, where $tau$ is the total time of the work-extraction protocol.
Our result highlights a new application of minimal dissipation processes and demonstrates a connection between thermodynamic geometry and the higher order statistical properties of work.
arXiv Detail & Related papers (2022-05-06T08:14:18Z) - Settling the Horizon-Dependence of Sample Complexity in Reinforcement
Learning [82.31436758872715]
We develop an algorithm that achieves the same PAC guarantee while using only $O(1)$ episodes of environment interactions.
We establish a connection between value functions in discounted and finite-horizon Markov decision processes.
arXiv Detail & Related papers (2021-11-01T00:21:24Z) - Simple scheme for extracting work with a single bath [0.0]
The protocol is based on a recent work definition involving only a single bath.
We quantify both the extracted work and the ideal efficiency of the process also giving maximum bounds for them.
Our proposal makes use of simple operations not needing fine control.
arXiv Detail & Related papers (2018-06-29T12:44:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.