Near-optimal protocols in complex nonequilibrium transformations
Edited by Ken A. Dill, Stony Brook University, Stony Brook, NY, and approved July 19, 2016 (received for review April 19, 2016)
Significance
Classical thermodynamics was developed to help design the best protocols for operating heat engines that remain close to equilibrium at all times. Modern experimental techniques for manipulating microscopic and mesoscopic systems routinely access far-from-equilibrium states, demanding new theoretical tools to describe the optimal protocols in this more complicated regime. Prior studies have sought, in simple models, the protocol that minimizes dissipation. We use computational tools to investigate the diversity of low-dissipation protocols. We show that optimal protocols can be accompanied by a vast set of near-optimal protocols, which still offer the substantive benefits of the optimal protocol. Although solving for the optimal protocol is typically difficult, computationally identifying a near-optimal protocol can be comparatively easy.
Abstract
The development of sophisticated experimental means to control nanoscale systems has motivated efforts to design driving protocols that minimize the energy dissipated to the environment. Computational models are a crucial tool in this practical challenge. We describe a general method for sampling an ensemble of finite-time, nonequilibrium protocols biased toward a low average dissipation. We show that this scheme can be carried out very efficiently in several limiting cases. As an application, we sample the ensemble of low-dissipation protocols that invert the magnetization of a 2D Ising model and explore how the diversity of the protocols varies in response to constraints on the average dissipation. In this example, we find that there is a large set of protocols with average dissipation close to the optimal value, which we argue is a general phenomenon.
Sign up for PNAS alerts.
Get alerts for new articles, or get an alert when an article is cited.
When a system is guided gradually from one equilibrium state to another, the amount of heat dissipated into its surroundings is insensitive to the manner of driving. In the more realistic case of an irreversible transformation in finite time, however, the dissipation can vary greatly from one driving protocol to another. These basic tenets of thermodynamics have received renewed attention in recent years due to improved capabilities for manipulating systems at small scales (1–7) and advances in the theoretical understanding of nonequilibrium fluctuations (8–10). In particular, many studies have sought to identify which finite-time protocols transform a system with the minimum amount of dissipation (11–18). Protocols that are optimal in this sense provide the most efficient route to measuring equilibrium free-energy differences—in simulations and in experiments (19)—via the Jarzynski relation (14, 20). More generally, low-dissipation protocols provide insight into the optimal design of nanoscale machines, both synthetic (4, 5) and natural (21).
However, it remains challenging to identify the minimum-dissipation protocol for complex, many-body systems driven far from equilibrium, despite recent progress (22–24). The difficulty of computing strictly optimal protocols motivates a pragmatic question: how large is the set of nearly optimal protocols? In this paper, we develop a framework to characterize that set. We introduce an entropy that indicates how many different protocols realize the same value of dissipation. For low values of dissipation, this protocol entropy quantifies how prevalent the near-optimal protocols are, highlighting when the system may be efficiently driven in many different ways. In analogy with common techniques of statistical physics, we present Monte Carlo methods to numerically compute the entropy by sampling protocols with a preference for low dissipation. The samples generated by this procedure demonstrate the distinct ways in which the system can be driven while maintaining the expectation of low dissipation. Variation among the sampled protocols accentuates features that are unimportant for ensuring low dissipation; similarly, the lack of variation highlights features that are essential for this goal. These ideas and capabilities complement previous approaches to determining optimal control procedures, using tools with many similar features (22, 24, 25). We elaborate on the connections in Discussion.
We illustrate our protocol-sampling framework with a numerical study of spin inversion of a ferromagnet, an essential process for copying information encoded in magnetic storage devices. Reducing dissipation in this context is of practical interest because thermodynamic costs of copying and erasing bits is projected to account for a significant fraction of future computational energy expenditures (26, 27). We examine a simple microscopic model of this process, based on the 2D Ising model (Fig. 1). At low temperature and in the presence of an external magnetic field, spins align strongly in the direction of the field. By adjusting the magnetic field and the temperature as functions of time, the magnetization may be rapidly inverted with a dissipation that depends on the manner in which the field and temperature are changed. Analysis of the protocol entropy in this model indicates that a large set of nonoptimal protocols can be used to control the system with a dissipation comparable to that of the optimal protocol.
Fig. 1.
Protocol Entropy
We first consider an ensemble of protocols sharing the same value ω of average dissipation. The protocol entropy of this ensemble measures the density of protocols that have mean dissipation ω. In analogy to the standard microcanonical ensemble of statistical mechanics, we define the entropywhere the integral runs over the space of time-dependent protocols and is a constant that sets the arbitrary zero of entropy. The delta function picks out protocols whose average dissipation lies within an infinitesimal interval around the specified value of ω. The subscript denotes an average taken over the probability distribution of stochastic trajectories evolving under the fixed protocol :For a single trajectory , the dissipation ω can be cast in terms of an imbalance between forward and reversed dynamics (10) givingTildes signify time reversal, so the numerator and denominator are probabilities of forward and reverse trajectories, respectively.
[1]
[2]
[3]
Optimal protocols, which carry some minimal dissipation, represent a small fraction of the possible protocols. Consequently, the protocol entropy evaluated at this minimal dissipation is low. As the dissipation increases, the entropy increases, and the growth rate reflects how flexibly low-dissipation protocols may be constructed. In particular, rapid entropic growth near the minimum dissipation suggests that targeting an exact optimal protocol is both challenging and gratuitous. Many other protocols, in practice, will perform comparably to the optimum.
The limiting behavior of near the minimum dissipation can be calculated exactly, assuming only that depends smoothly on variations in the low-dissipation protocols. As shown in Supporting Information, protocol entropy grows logarithmically in this regime:where n is the total number of protocol degrees of freedom, that is, the total number of parameter values defining any given protocol. Quite generally, therefore, protocol entropy increases sharply over a narrow range of dissipation values just above the minimum
[4]
A Canonical Protocol Ensemble
To numerically compute the protocol entropy, it is useful to introduce a canonical protocol ensemble with distributionIn correspondence with the canonical ensemble of statistical mechanics, acts as an effective energy for each protocol and γ plays the role of inverse temperature, tuning the mean dissipation, that is, the average of ω over the distribution . Searching for a protocol with strictly minimum dissipation amounts to a zero “temperature” () quench, whereas near-optimal protocols are identified by large values of γ.
[5]
When using a sufficiently large γ, the samples reveal representative low-dissipation protocols. By using different biasing strengths γ in Eq. 5, we can learn about the characteristics of protocols with distinct values of average dissipation. In addition, the protocols sampled with various choices of γ can be combined to calculate over a broad range.
Sampling Protocols and Trajectories.
In principle, the ensemble defined by Eq. 5 may be directly sampled with a Monte Carlo procedure that conditionally accepts protocol changes based on the corresponding changes in . For complex systems, however, values of are typically not known exactly. They can be estimated from the sample mean of a collection of N trajectories drawn from . However, for finite N, replacing by in Eq. 5 yields a distribution of protocols that differs from . Strategies to correct for the finite-N bias have been formulated to enable conventional Boltzmann sampling when configurational energies cannot be calculated with certainty (28–32). Here, we consider an analogous strategy in the context of sampling protocols.
To sample the canonical protocol distribution, we construct a Monte Carlo procedure that performs a random walk through the joint space of protocols and N independent trajectories . A trial move amounts to an attempt to make changes in both and in . Operationally, this proposed change can be achieved by first perturbing the protocol (with a symmetric generation probability) before regenerating the trajectories using the new protocol. For simplicity, we consider the case that the trajectory generation probabilities are also symmetric, as in the noise-guided shooting procedures of transition path sampling (33–36). Accepting a trial move with the Metropolis probability , where is the difference between the sample means under the original and trial protocols, yields the stationary distributionThe resulting marginal distribution of protocols,is determined by the dissipation statistics of each protocol, but in a more complicated way than . Nevertheless, we show that the sampled protocols are drawn from a canonical protocol distribution in two special situations: the case of Gaussian dissipation distributions and the large N limit.
[6]
[7]
Cumulant Expansion.
The expression for the marginal distribution Eq. 7 may be recast aswhere is the cumulant-generating function for the dissipation. A cumulant expansion,relates the probability of sampled protocols to the cumulants of the dissipation distribution , where . As N grows, the contribution from higher-order cumulants vanishes, consistent with the central limit theorem. In the limit , sampling Eq. 5 becomes equivalent to sampling Eq. 6 because the sample mean converges to the average dissipation.
[8]
[9]
In the special case that is Gaussian for each protocol , a powerful simplification arises, averting the need to use large values of N. Gaussian dissipation distributions occur in many contexts—as a defining feature of linear response (37), in the limit of slow adiabatic driving (37), and when Brownian particles evolve in driven harmonic potentials (38). In all these cases, the cumulants of beyond the variance vanish, allowing us to exactly truncate Eq. 9 at second order. If we further take to be symmetric under time reversal, then the fluctuation theorem provides an exact relationship between the mean and variance: . As a result, the biased protocol distribution can be expressed in terms of mean dissipation alone:Eq. 10 has precisely the form of the canonical protocol distribution (Eq. 5), with an effective bias . This result offers tremendous flexibility. An exact bias toward low average dissipation can be achieved with any N, for example, by sampling a small number of trajectories for each proposed change in protocol. Because generating trajectories dominates the computational expense of our sampling scheme, the freedom to choose small N is very attractive.
[10]
The limitation with using small N when sampling protocols is that the achievable bias strength γ cannot exceed . This constraint arises because the λ bias in Eq. 5 directly favors low-dissipation trajectories, and not necessarily low-dissipation protocols. For small values of λ, the trajectories sampled for a given protocol are typical of the unbiased trajectory distribution . In this case, there is thus a strong correlation between sampled low-dissipation trajectories and protocols that yield low dissipation on average. This correspondence degrades for large λ. In fact, for , sampled trajectories have negative dissipation on average,† which cannot be typical of any protocol according to the second law of thermodynamics. Thus, as λ is increased toward , the joint ensemble of trajectories and protocols switches from highlighting low-dissipation protocols to emphasizing rare negative-dissipation trajectories. Moreover, sampling with large values of λ requires generation of increasingly rare trajectories, complicating efficient path sampling as discussed in Supporting Information.
Results
Spin Inversion Protocols.
To illustrate the use of the low-dissipation protocol ensemble, we consider the inversion of spins in a ferromagnet. Specifically, we imagine initializing a system of interacting spins at low temperature, where its equilibrium state has long-range “up” or “down” order. We then ask how best to flip this “bit.” That is, how should we vary the temperature and external field as functions of time to flip the state of the magnet without excess dissipation? This problem, relevant to the design of low-power magnetic hard drives, has been investigated as an optimal control problem elsewhere (23, 40). Here, we also consider the near-optimal drivings.
We represent the ferromagnet as a 2D Ising model with periodic boundary conditions and dynamics generated by a succession of individual spin flips. With an external magnetic field h, the energy of a configuration is given bywhere , and indicates a sum over nearest neighbor sites i and j. An attempted spin flip that alters the energy by is accepted with Glauber probability , where T is the temperature of the bath. Unlike equilibrium Ising model dynamics, the temperature and magnetic field are time dependent as prescribed by a nonequilibrium protocol . In a finite amount of time , we aim to switch from the macroscopic up state to the down state. We therefore consider only protocols that begin and end at low temperature and that switch from a positive to a negative field .
[11]
One consequence of the nonequilibrium driving is that the dynamics is not microscopically reversible. For ordinary Glauber dynamics, the equilibrium probability of a trajectory segment is equal to its time-reversed counterpart, but our time-dependent driving breaks this equality. By tracking the random numbers that generate each spin flip, we explicitly compute the forward and time-reversed probabilities of each trajectory, thereby computing the stochastic thermodynamic dissipation via Eq. 3. Physically, the dissipation of each microscopic step multiplied by T is the heat transferred from the thermal bath into the system.
We use Monte Carlo techniques, discussed further in Materials and Methods, to sample low-dissipation protocols from Eq. 6 with . Fig. 1 shows 450 representative protocols, all of which avoid the region of parameter space near the Ising critical point. Control is particularly costly (23) in this vicinity due to critical slowing down, which causes the spin system to lag behind changes in the control parameters. There is a natural connection between dissipation and lag (41, 42): the farther the system falls out of equilibrium with control parameters’ instantaneous values, the more heat is dissipated to the reservoir during the relaxation.
Roughly, the optimal strategy requires that we first heat the magnet, next invert the field, and then cool the magnet. However, the varied protocols in Fig. 1 demonstrate significant leeway in how these steps are carried out. Most notably, while the system is held at low temperature, the magnetic field need not be precisely tuned, as evidenced by large variations both early and late in the protocol. Some low-dissipation protocols even transiently invert the field at low temperature, thereby crossing the equilibrium coexistence curve, only to restore the field’s original sign a short time later. This seemingly wasteful procedure in fact incurs little dissipative cost, because it is highly ineffectual. The low-temperature field inversion is too brief for nucleation of the new phase to occur with significant probability, so the extent of relaxation is negligible. Absent relaxation, no heat is dissipated to the bath. Perhaps counterintuitively, the lag is so severe in this case as to be irrelevant.
Protocol Entropy.
The protocol entropy determined by sampling magnetization inversion dynamics is shown in Fig. 2. Near the apparent minimum dissipation, a small increase in mean dissipation is accompanied by a steep rise in , that is, the number of protocols grows rapidly as we permit modest excess dissipation. In fact, the density of protocols increases by several orders of magnitude for a change in dissipation that is very small relative to dissipation fluctuations for a fixed protocol. This rapid initial growth is captured well by the asymptotic form of Eq. 4, which depends on the number of degrees of freedom in the protocol. Farther from the minimal value of mean dissipation, the protocol entropy climbs much more gradually.
Fig. 2.
The slope of reflects the strength of bias γ needed to depress the average dissipation. The Inset of Fig. 2 illustrates a crossover between two regimes: small biases greatly reduce the mean dissipation but further reduction requires very large biases. Thus, weak biases on can be greatly effective at directing the protocol sampling toward the optimum. We anticipate that these limiting behaviors are generic, and the corresponding asymptotic forms are derived in Supporting Information. The reduction in dissipation due to small values of γ is governed by the variability of in an unbiased protocol ensemble. Because complex systems typically depend sensitively on one or more of their control parameters, this variability should be substantial in general. Large values of γ favor nearly optimal protocols, whose diversity is well described by Eq. 4. Correspondingly, the mean dissipation in the large γ limit decays slowly as
Gaussian Fluctuations.
We have computed for the spin inversion process both with and without the simplifying assumption that the dissipation is Gaussian distributed. We find that the Gaussian approximation provides an estimate that is accurate within statistical error despite requiring a significantly reduced number of trajectories. To more explicitly demonstrate the validity of the approximation, we selected three protocols from our sampling, which are shown in Fig. 3A.
Fig. 3.
For each protocol, we computed the dissipation distribution . Empirically, we find that these distributions are strikingly Gaussian over a large range of ω that includes . At large positive values of dissipation, we observe “fat” exponential tails, consistent with the structure of generic current large deviation functions that has recently been demonstrated for the case of time-independent driving (44, 45). This fat tail, associated with clusters of spins that resist reorientation, only weakly restricts our use of the Gaussian dissipation assumption. The positive λ bias, useful to study low-dissipation behavior, focuses the sampling toward the Gaussian region of the distribution, rendering the exponential tails insignificant.
Discussion
Low dissipation, the focus of our exploration of driving protocols, is one of many possible objectives for nonequilibrium control. Indeed, minimizing dissipation can be viewed as an instance of the extensively studied problem of stochastic optimal control (46–50). The formulation of stochastic optimal control as a path integral control problem (25, 51–53) has a particularly close connection to our work, in that importance sampling of trajectories can be used to iteratively refine a protocol toward the optimum (22, 24). A distinct feature of our approach is the systematic sampling of protocols, as opposed to the more limited goal of strictly optimizing them. By simultaneously sampling trajectories and protocols with a well-defined bias, we identify low-dissipation protocols without the challenging task of converging trajectory-space averages for any particular protocol, as is required in iterative optimization methods. Protocols harvested by our procedure, in contrast to those encountered in an optimization, quantitatively reflect the diversity of low-dissipation protocols.
Scrutinizing near-optimal protocols complements the search for optimality in several ways. In simple model systems that can be optimized exactly, minimum-dissipation protocols are known to involve features that are singular or may be impractical to implement (14). In such cases, the collection of near-optimal protocols becomes a natural target for design. Even when the optimal protocol may be physically achieved, its form does not directly indicate which features of a driving history are essential to its success, and which are irrelevant. In our approach, relevant features can be readily identified through their limited variability in the protocol ensemble, as illustrated by the cloud of protocols in Fig. 1B.
Finally, we note that efficient but suboptimal nonequilibrium transformations are almost certainly the norm in biology at many scales. Indeed, the evolutionary dynamics of biological adaptation might be viewed as an importance sampling on the space of protocols, roughly akin to the sampling methods developed in this paper. The surprising, often eccentric strategies used to perform simple tasks in biology are, perhaps, indicative of the myriad options provided by an ensemble of protocols evolving under a complex set of constraints.
Materials and Methods
We sample the joint space of trajectories and protocols using Markov chain Monte Carlo methods. Each point in this space consists of a protocol and N independent trajectories subject to that protocol. With tunable bias λ, the Markov chain samples the distribution in Eq. 6. We restrict the space of protocols with a view toward experimental practicality. For the 2D Ising example, we impose two such restrictions. First, to allow only slowly varying protocols, we parameterize our protocol space by the values of temperature and external field at 11 evenly spaced times. We call the values at these special times the control points. Between any two neighboring control points, the temperature and field strength are linearly interpolated. Second, we require that and for all control points to mimic that physical apparatuses can tune controls over bounded ranges. Limiting the protocol space in these ways can be viewed as a regularization scheme that simplifies the representation problem of optimal control theory (24). We note, however, that the protocol entropy depends on our choice of regularization. If, for example, we were to use many more control points with the same , then we would introduce many additional protocols with high-frequency features. Consequently, would grow more rapidly, following Eq. 4. Fig. S1 highlights this dependence on the number of protocol degrees of freedom.
Fig. S1.
Each Monte Carlo move first attempts to adjust the protocol by moving a single control point by a random displacement in the temperature–field plane. The move is constructed to be symmetric, meaning the probability of selecting any displacement vector equals the probability of a displacement of the opposite sign. Using this trial protocol, N new trajectories are simulated using a sequence of Glauber single spin flips. Conventionally, each step of the Glauber Ising dynamics chooses a random spin, which is flipped to generate a trial configuration. To enable more efficient noise-guided trajectory sampling, we use a modified Glauber dynamics: the trial configuration is given by setting the randomly selected spin to either the up or the down state without reference to its prior state (36). The move is futile when the selected spin is already in the trial configuration. Because one-half of the moves are futile on average, the Monte Carlo time is rescaled by a factor of 2 compared with ordinary single-spin flip Glauber dynamics. Following each move, the probability of running that step backward is computed, enabling an explicit calculation of the dissipation of each trajectory. The new protocol and trajectories are conditionally accepted with probability , where is the difference between the sample means under the original and trial protocols.
The protocol entropy is calculated, up to a constant offset , using a weighted average over the protocols collected by the Monte Carlo procedure:
[12]
From a set of M sampled protocols, , we therefore estimate the following:
[13]
Operationally, this amounts to collecting a histogram of values of with each entry weighted by the corresponding value of . To generate Fig. 2, each of these weights is computed by estimating the exponential average from 1,000 independent trajectories. In practice, is constructed using the multistate Bennett acceptance ratio (MBAR) method to combine samples collected with and with several different values of the bias ranging from to . The offset is chosen such that is zero at its maximum.
The protocol entropy can be computed much more efficiently when the Gaussian approximation may be used to evaluate the exponential average. To evaluate the validity of this approximation for the Ising dynamics, we compute the actual dissipation distributions by sampling trajectories with a fixed protocol. These trajectories are importance sampled using harmonic biases, which restrain the dissipation to fluctuate around a specified value. By choosing several different harmonic biases, trajectories were biased into both tails of the dissipation distribution, which was reconstructing using the MBAR method (43).
When the Gaussian approximation is appropriate, it is wasteful to use a large value of N. Low-dissipation protocols may be sampled with an effective biasing strength using various combinations of N and λ, and a small N reduces the computational expense, as demonstrated in Fig. S2. However, when N is too small or λ too large, the Monte Carlo acceptance probability drops precipitously, a fact elaborated upon in Supporting Information, particularly in Fig. S3. Sampling efficiency is poor under these conditions because the Markov chain favors a collection of rare trajectories with dissipation below the mean (and often below zero). This issue can be partially alleviated by introducing replica exchange moves—random swaps exchanging replicas with different biasing strengths λ. The implementation of this procedure naturally mirrors the use of replica exchange to surmount kinetic traps when sampling low-temperature molecular configurations. Further performance enhancements are obtained when trial trajectories are generated with random numbers (noises) that correlate with the noises of the previous collection of trajectories. An implementation of this noise-guided sampling is described in detail elsewhere (36, 54). The noise guidance technique is not strictly required to perform the protocol sampling, but in Fig. S4 we show that it can provide significant practical benefits. Fig. S5 shows further practical improvements.
Fig. S2.
Fig. S3.
Fig. S4.
Fig. S5.
General Features of the Protocol Entropy
In the limit of large values of the bias parameter the difference between a sampled protocol and the optimal protocol should be small. If the average dissipation is a smooth functional of the protocol, we can approximate the deviation from the minimum average dissipation in terms of the small protocol variations :where the indices i and j label times at which the protocol is manipulated. In this limit, we can explicitly calculate the moment generating function for the average dissipation:and n denotes the total number of degrees of freedom in the protocol.
[S1]
[S2]
The unbiased distribution of average dissipations is an inverse Laplace transform of and hence the entropy in the large γ limit isFig. S1 compares the asymptotic expression (Eq. S4) with the entropy computed for the Ising system.
[S3]
[S4]
From the asymptotic expression for , we can also compute the average dissipation associated with protocols sampled under the bias γ:In the absence of protocol bias, , typical values of dissipation are quite large. Provided that is bounded, the mean dissipation at should nonetheless be finite, as is the corresponding variance . Sufficiently close to the maximum of , we therefore have the following:The parameters and in this expression are determined numerically by computing average dissipation for protocols generated in the ensemble. The corresponding curve is plotted in Fig. 2 as a dotted line.
[S5]
[S6]
[S7]
In the small-bias regime, the moment generating function,gives a γ-biased average dissipation:which is shown as a dotted line in the Inset of Fig. 2.
[S8]
[S9]
Sampling Efficiency
Ergodic sampling of requires that decorrelated protocols be generated, and the efficiency of the sampling depends on how many Monte Carlo moves are required to produce each decorrelated sample. As discussed in the main text, the canonical distribution with bias γ may be accessed with various choices of N and λ, yet the sampling efficiency depends on these choices. A large value of N bears a clear computational cost because N new trajectories must be simulated for each trial protocol. It is not, however, always optimal to choose the minimal N capable of generating bias γ, . The problem with using a small value of N is that rare low-dissipation trajectories (including those with negative dissipation) can depress the probability of accepting a protocol move, even when the trial protocol has a lower average dissipation. The smaller the choice of N, the more rare trajectories influence the acceptance of protocol moves. Consequently, the optimal choice of N often exceeds .
In practice, the best choice of N depends on the Monte Carlo moves used to update trajectories and protocols. For example, there is a trade-off in choosing the best protocol space moves. Large changes to the protocol have a low acceptance probability, but small steps require many moves before sampling a protocol with a decorrelated value of mean dissipation. To quantify the trade-offs between possible choices of N and of Monte Carlo moves, we construct a correlation function:where is the difference between the ith sampled protocol’s average dissipation and the mean dissipation, found by averaging over all protocols in the ensemble. Assuming exponentially decaying correlations (), we set the correlation time to the decay constant . This decorrelation time is the typical number of protocols that must be sampled before generating a protocol with a decorrelated value of the mean dissipation. Because each new protocol must be accompanied by the simulation of N new trajectories, the computational cost for obtaining each decorrelated sample is given by .
[S10]
The decorrelation time implicitly depends on N as well as on the details of the sampling moves. To gain intuition about the most computationally efficient choice of N, we first consider a simplified Gaussian model. For this model, the optimal N slightly exceeds . The Ising dynamics is more complicated, requiring to be computed from simulations. We show that the computational expense depends on N in a similar manner as in the Gaussian model, although the computational expense may be reduced by using noise-guided path sampling.
Gaussian Model.
We sample a scalar protocol in the vicinity of the minimum dissipation protocol . The advantage of sampling near is that we can Taylor expand to arrive at an expression for the average dissipation associated with a perturbation :We assume that the dissipation distribution associated with a protocol is Gaussian:We furthermore use a Gaussian distribution to propose a trial protocol:Using the new trial protocol, the dissipation for N independent trajectories is drawn from Eq. S12 to compute the sample mean dissipation for the new protocol . For a choice of bias γ, we choose λ and N such that and accept the new protocol and trajectories with the following probability:where is the difference between the new protocol’s sample mean dissipation and that of the old protocol and trajectories.
[S11]
[S12]
[S13]
[S14]
The correlation between subsequent samples is given bywithgiving the acceptance probability for the Monte Carlo protocol move. Because the sample mean dissipations (for both the old and trial protocol) are Gaussian distributed, is also a Gaussian:The acceptance rate may therefore be expressed in terms of error functions. The number of attempted protocol moves required to sample protocols with decorrelated mean dissipations is given by . A complicated integral expression for in terms of and N may be derived in terms of error functions. Numerically evaluating the integral expression yields the N dependence of the computational cost, , for generating protocols with decorrelated average dissipation. Analogously, we may compute the time to find decorrelated values of the scalar protocol as the decay constant for the correlation function:Both and yield the same heuristic: the computational expense is minimal when N slightly exceeds , as plotted in Fig. S2.
[S15]
[S16]
[S17]
[S18]
Ising Dynamics.
Using various choices of λ and N, the canonical protocol ensemble is sampled by a sequence of protocol moves that alter the control point at a single time. Either the control point’s magnetic field strength is increased by a random uniform number between and 0.5 or the temperature is increased by a random uniform number between and 2.5. The protocol move is resampled if the magnitude of the magnetic field strength at any time exceeds 1 or if the dimensionless temperature exits the range . N trajectories are generated by repropagating forward dynamics from the initial time or by running time-reversed dynamics from the final time slice and reweighting the trajectories into the forward-trajectory ensemble. When noise guidance path sampling is not being used, each trial trajectory is generated by drawing new random noises to carry out the steps of the spin flip dynamics. The noise guidance path sampling scheme recycles, with probability , each random number from the previous trajectory. Over time, each random number gets resampled uniformly, but the correlations between an old trajectory and a new trajectory are enhanced (36, 54).
To assess how rapidly the protocol space is sampled, protocols were stored and an accurate estimate of was found by averaging over 1,000 trajectories with the fixed protocol. Fig. S3A shows the (Monte Carlo) time series of this average dissipation. Using this time series, the correlation function is computed for each choice of N and λ. The correlation functions are fit to exponential decays, and the decay constant τ is extracted to yield the computational cost as a function of N. Fig. S4 illustrates that the optimal choice of N exceeds , as in the Gaussian model. The noise guidance strategy aids small-N sampling, thereby lowering the optimal N. For these small choices of N, the noise guidance scheme offers a roughly order of magnitude speed up.
Sample Mean Fluctuations
The biasing strength γ necessary to sample protocols with dissipation is given by the slope of at . When is especially steep, it therefore requires a very strong bias to access the low-dissipation entropy. Sampling efficiency worsens for large γ, so we note an alternative method for computing that makes use of the fluctuations in the sample mean of N trajectories.
We define the entropy to give the density of protocols and trajectories that have a sample mean dissipation within an infinitesimal window of ω:In the limit , this entropy must converge to the protocol entropy, , but falls off less rapidly around the minimal average dissipation. The tails of with low sample mean may therefore be importance sampled with a bias Sampling directly, on the other hand, requires the stronger bias γ. By making the Gaussian approximation, we can reconstruct from N dependence of using the following:a technique illustrated in Fig. S5.
[S19]
[S20]
Acknowledgments
T.R.G. acknowledges support from the National Science Foundation (NSF) Graduate Research Fellowship, the Fannie and John Hertz Foundation, and the Gordon and Betty Moore Foundation as a Massachusetts Institute of Technology Physics of Living Systems Fellow. G.M.R. acknowledges support from the NSF Graduate Research Fellowship. P.L.G. was supported by the US Department of Energy, Office of Basic Energy Sciences, through the Chemical Sciences Division of the Lawrence Berkeley National Laboratory, under Contract DE-AC02-05CH11231. G.E.C. acknowledges support from the US Army Research Laboratory and the US Army Research Office under Contract W911NF-13-1-0390.
Supporting Information
Supporting Information (PDF)
Supporting Information
- Download
- 480.86 KB
References
1
SB Smith, L Finzi, C Bustamante, Direct mechanical measurements of the elasticity of single DNA molecules by using magnetic beads. Science 258, 1122–1126 (1992).
2
J Liphardt, S Dumont, SB Smith, Jr I Tinoco, C Bustamante, Equilibrium information from nonequilibrium measurements in an experimental test of Jarzynski’s equality. Science 296, 1832–1835 (2002).
3
D Collin, et al., Verification of the Crooks fluctuation theorem and recovery of RNA folding free energies. Nature 437, 231–234 (2005).
4
IA Martínez, et al., Brownian Carnot engine. Nat Phys 12, 67–70 (2016).
5
V Blickle, C Bechinger, Realization of a micrometre-sized stochastic heat engine. Nat Phys 8, 143–146 (2012).
6
S Toyabe, T Sagawa, M Ueda, E Muneyuki, M Sano, Experimental demonstration of information-to-energy conversion and validation of the generalized Jarzynski equality. Nat Phys 6, 988–992 (2010).
7
Y Jun, M Gavrilov, J Bechhoefer, High-precision test of Landauer’s principle in a feedback trap. Phys Rev Lett 113, 190601 (2014).
8
C Jarzynski, Nonequilibrium equality for free energy differences. Phys Rev Lett 78, 2690 (1997).
9
GE Crooks, Excursions in statistical dynamics. PhD thesis (University of California, Berkeley). (1999).
10
R Spinney, I Ford, Fluctuation relations: A pedagogical overview. Nonequilibrium Statistical Physics of Small Systems: Fluctuation Relations and Beyond, eds R Klages, W Just, C Jarzynski (Wiley-VCH, Weinheim, Germany), pp. 3–56 (2013).
11
F Weinhold, Metric geometry of equilibrium thermodynamics. J Chem Phys 63, 2479–2483 (1975).
12
G Ruppeiner, Thermodynamics: A Riemannian geometric model. Phys Rev A 20, 1608–1613 (1979).
13
P Salamon, A Nitzan, B Andresen, RS Berry, Minimum entropy production and the optimization of heat engines. Phys Rev A 21, 2115–2129 (1980).
14
T Schmiedl, U Seifert, Optimal finite-time processes in stochastic thermodynamics. Phys Rev Lett 98, 108301 (2007).
15
A Gomez-Marin, T Schmiedl, U Seifert, Optimal protocols for minimal work processes in underdamped stochastic thermodynamics. J Chem Phys 129, 024114 (2008).
16
H Then, A Engel, Computing the optimal protocol for finite-time processes in stochastic thermodynamics. Phys Rev E Stat Nonlin Soft Matter Phys 77, 041105 (2008).
17
PR Zulkowski, DA Sivak, GE Crooks, MR DeWeese, Geometry of thermodynamic control. Phys Rev E Stat Nonlin Soft Matter Phys 86, 041148 (2012).
18
PR Zulkowski, MR DeWeese, Optimal control of overdamped systems. Phys Rev E Stat Nonlin Soft Matter Phys 92, 032117 (2015).
19
P Maragakis, F Ritort, C Bustamante, M Karplus, GE Crooks, Bayesian estimates of free energies from nonequilibrium work data in the presence of instrument noise. J Chem Phys 129, 024102 (2008).
20
C Dellago, G Hummer, Computing equilibrium free energies using non-equilibrium molecular dynamics. Entropy (Basel) 16, 41–61 (2013).
21
G Oster, H Wang, Why is the mechanical efficiency of F1-ATPase so high? J Bioenerg Biomembr 32, 459–469 (2000).
22
W Zhang, H Wang, C Hartmann, M Weber, C Schütte, Applications of the cross-entropy method to importance sampling and optimal control of diffusions. SIAM J Sci Comput 36, A2654–A2672 (2014).
23
GM Rotskoff, GE Crooks, Optimal control in nonequilibrium systems: Dynamic Riemannian geometry of the Ising model. Phys Rev E Stat Nonlin Soft Matter Phys 92, 060102 (2015).
24
HJ Kappen, HC Ruiz, Adaptive importance sampling for control and inference. J Stat Phys 162, 1244–1266 (2016).
25
HJ Kappen, Linear theory for control of nonlinear stochastic systems. Phys Rev Lett 95, 200201 (2005).
26
B Lambson, D Carlton, J Bokor, Exploring the thermodynamic limits of computation in integrated systems: Magnetic memory, nanomagnetic logic, and the Landauer limit. Phys Rev Lett 107, 010604 (2011).
27
J Hong, B Lambson, S Dhuey, J Bokor, Experimental test of Landauer’s principle in single-bit operations on nanomagnetic memory bits. Sci Adv 2, e1501492 (2016).
28
C Andrieu, GO Roberts, The pseudo-marginal approach for efficient Monte Carlo computations. Ann Stat 37, 697–725 (2009).
29
RC Ball, TMA Fink, NE Bowler, Stochastic annealing. Phys Rev Lett 91, 030201 (2003).
30
MA Beaumont, Estimation of population growth or decline in genetically monitored populations. Genetics 164, 1139–1160 (2003).
31
DM Ceperley, M Dewing, The penalty method for random walks with uncertain energies. J Chem Phys 110, 9812–9820 (1999).
32
L Lin, KF Liu, J Sloan, A noisy Monte Carlo algorithm. Phys Rev D Part Fields 61, 074505 (2000).
33
C Dellago, PG Bolhuis, D Chandler, Efficient transition path sampling: Application to Lennard-Jones cluster rearrangements. J Chem Phys 108, 9236–9245 (1998).
34
GE Crooks, D Chandler, Efficient transition path sampling for nonequilibrium stochastic dynamics. Phys Rev E Stat Nonlin Soft Matter Phys 64, 026109 (2001).
35
AK Hartmann, High-precision work distributions for extreme nonequilibrium processes in large systems. Phys Rev E Stat Nonlin Soft Matter Phys 89, 052103 (2014).
36
TR Gingrich, PL Geissler, Preserving correlations between trajectories for efficient path sampling. J Chem Phys 142, 234104 (2015).
37
T Speck, U Seifert, Distribution of work in isothermal nonequilibrium processes. Phys Rev E Stat Nonlin Soft Matter Phys 70, 066112 (2004).
38
O Mazonka, C Jarzynski, Exactly solvable model illustrating far-from-equilibrium predictions. arXiv:cond-mat/9912121. (1999).
39
JL Lebowitz, H Spohn, A Gallavotti–Cohen-type symmetry in the large deviation functional for stochastic dynamics. J Stat Phys 95, 333–365 (1999).
40
M Venturoli, E Vanden-Eijnden, G Ciccotti, Kinetics of phase transitions in two dimensional Ising models studied with the string method. J Math Chem 45, 188–222 (2009).
41
R Kawai, JMR Parrondo, C Van den Broeck, Dissipation: The phase-space perspective. Phys Rev Lett 98, 080602 (2007).
42
S Vaikuntanathan, C Jarzynski, Dissipation and lag in irreversible processes. EPL 87, 60005 (2009).
43
MR Shirts, JD Chodera, Statistically optimal analysis of samples from multiple equilibrium states. J Chem Phys 129, 124105 (2008).
44
TR Gingrich, JM Horowitz, N Perunov, JL England, Dissipation bounds all steady-state current fluctuations. Phys Rev Lett 116, 120601 (2016).
45
P Pietzonka, AC Barato, U Seifert, Universal bounds on current fluctuations. Phys Rev E Stat Nonlin Soft Matter Phys 93, 052145 (2016).
46
WH Fleming, Stochastic control for small noise intensities. SIAM J Control 9, 473–517 (1971).
47
WH Fleming, Exit probabilities and optimal stochastic control. Appl Math Optim 4, 329–346 (1977).
48
P Dupuis, H Wang, Importance sampling, large deviations, and differential games. Stochastics 76, 481–508 (2004).
49
C Hartmann, C Schütte, Efficient rare event simulation by optimal nonequilibrium forcing. J Stat Mech 2012, P11004 (2012).
50
R Chetrite, H Touchette, Variational and optimal control representations of conditioned and driven processes. J Stat Mech 2015, P12001 (2015).
51
E Theodorou, J Buchli, S Schaal, A generalized path integral control approach to reinforcement learning. J Mach Learn Res 11, 3137–3181 (2010).
52
VY Chernyak, M Chertkov, J Bierkens, HJ Kappen, Stochastic optimal control as non-equilibrium statistical mechanics: Calculus of variations over density and current. J Phys A 47, 022001 (2014).
53
S Thijssen, HJ Kappen, Path integral control and state-dependent feedback. Phys Rev E Stat Nonlin Soft Matter Phys 91, 032104 (2015).
54
TR Gingrich, Two paths diverged: Exploring trajectories, protocols, and dynamic phases. PhD thesis (University of California, Berkeley). (2015).
Information & Authors
Information
Published in
Classifications
Submission history
Published online: August 29, 2016
Published in issue: September 13, 2016
Keywords
Acknowledgments
T.R.G. acknowledges support from the National Science Foundation (NSF) Graduate Research Fellowship, the Fannie and John Hertz Foundation, and the Gordon and Betty Moore Foundation as a Massachusetts Institute of Technology Physics of Living Systems Fellow. G.M.R. acknowledges support from the NSF Graduate Research Fellowship. P.L.G. was supported by the US Department of Energy, Office of Basic Energy Sciences, through the Chemical Sciences Division of the Lawrence Berkeley National Laboratory, under Contract DE-AC02-05CH11231. G.E.C. acknowledges support from the US Army Research Laboratory and the US Army Research Office under Contract W911NF-13-1-0390.
Notes
This article is a PNAS Direct Submission.
†
The cumulant-generating function is symmetric about (39), and is correspondingly symmetric in λ about . This symmetry implies negative average dissipation when .
Authors
Competing Interests
The authors declare no conflict of interest.
Metrics & Citations
Metrics
Citation statements
Altmetrics
Citations
Cite this article
113 (37) 10263-10268,
Export the article citation data by selecting a format from the list below and clicking Export.
Cited by
Loading...
View Options
View options
PDF format
Download this article as a PDF file
DOWNLOAD PDFLogin options
Check if you have access through your login credentials or your institution to get full access on this article.
Personal login Institutional LoginRecommend to a librarian
Recommend PNAS to a LibrarianPurchase options
Purchase this article to access the full text.