## New Research In

### Physical Sciences

### Social Sciences

#### Featured Portals

#### Articles by Topic

### Biological Sciences

#### Featured Portals

#### Articles by Topic

- Agricultural Sciences
- Anthropology
- Applied Biological Sciences
- Biochemistry
- Biophysics and Computational Biology
- Cell Biology
- Developmental Biology
- Ecology
- Environmental Sciences
- Evolution
- Genetics
- Immunology and Inflammation
- Medical Sciences
- Microbiology
- Neuroscience
- Pharmacology
- Physiology
- Plant Biology
- Population Biology
- Psychological and Cognitive Sciences
- Sustainability Science
- Systems Biology

# Energetic costs of cellular computation

Edited by Herbert Levine, University of California, San Diego, La Jolla, CA, and approved September 10, 2012 (received for review May 9, 2012)

## Abstract

Cells often perform computations in order to respond to environmental cues. A simple example is the classic problem, first considered by Berg and Purcell, of determining the concentration of a chemical ligand in the surrounding media. On general theoretical grounds, it is expected that such computations require cells to consume energy. In particular, Landauer’s principle states that energy must be consumed in order to erase the memory of past observations. Here, we explicitly calculate the energetic cost of steady-state computation of ligand concentration for a simple two-component cellular network that implements a noisy version of the Berg–Purcell strategy. We show that learning about external concentrations necessitates the breaking of detailed balance and consumption of energy, with greater learning requiring more energy. Our calculations suggest that the energetic costs of cellular computation may be an important constraint on networks designed to function in resource poor environments, such as the spore germination networks of bacteria.

The relationship between information and thermodynamics remains an active area of research despite decades of study (1⇓⇓–4). An important implication of the recent experimental confirmation of Landauer’s principle, relating the erasure of information to thermodynamic irreversibility, is that any irreversible computing device must necessarily consume energy (2, 3). The generality of Landauer’s argument suggests that it is true regardless of how the computation is implemented. A particularly interesting class of examples relevant to systems biology and biophysics is that of intracellular biochemical networks that compute information about the external environment. These biochemical networks are ubiquitous in biology, ranging from the quorum-sensing and chemotaxis networks in single-cell organisms to networks that detect hormones and other signaling factors in higher organisms.

A fundamental issue is the relationship between the information processing capabilities of these biochemical networks and their energetic costs (5⇓⇓–8). It is known that energetic costs place important constraints on the design of physical computing devices as well as on neural computing architectures in the brain and retina (9⇓–11), suggesting that these constraints may also influence the design of cellular computing networks.

The best studied example of a cellular computation is the estimation of the steady-state concentration of a chemical ligand in the surrounding environment (12⇓–14). This problem was first considered in the seminal paper by Berg and Purcell who showed that the information a cell can acquire about its environment is fundamentally limited by stochastic fluctuations in the occupancy of the membrane-bound receptor proteins that detect the ligand (12). In particular, they considered the case of a cellular receptor that binds ligands with a concentration-dependent rate and unbinds particles at a uniform rate (see Fig. 1). They argued that cells could estimate the ambient chemical concentration by calculating the average time a receptor is bound during a chosen measurement time *T*≫1. Recently, however, it was shown that the optimal strategy for a cell is instead to calculate the average duration of the unbound intervals during *T* or, equivalently, the total time that the receptor was unbound during *T*. This later computation implements maximum likelihood estimation (MLE) (14). In these previous studies, the biochemical network downstream of the receptors that implements the desired computations was largely ignored because the authors were primarily interested in calculating fundamental limits on how precisely cells can compute external concentrations. However, calculating energetic costs requires us to explicitly model the downstream biochemical networks that implement these computations (15).

Here, we consider a simple two-component biochemical network that encodes information about ligand concentration in the steady-state concentration of the activated form of a downstream protein (as shown in Fig. 1). Such two-component networks are a common signal transduction motif found in bacteria and are often used to sense external signals through receptor-catalyzed phosphorylation of a downstream response regulator (16). The membrane-bound receptors can be in an either an active “on” state or an inactive “off” state. For simplicity, as in previous works (12⇓–14), we assume that the binding affinity of the on state is extremely high such that all ligand-bound receptors are always in the on state and all unbound receptors are in the off state. Receptors can switch between the off state and on state at a concentration-dependent rate and from the on state to the off state at a concentration-independent rate . Receptors additionally convert a downstream signaling protein from an inactive form *X* to an active form *X*^{∗}, by, for example, phosphorylation, at a state-dependent rate , where *s* = on,off. The proteins are deactivated at a state-independent rate *k*_{1}. The dependence of on the receptor state is what propagates information about ligand concentration downstream. The biochemical network described above contains all the basic elements of a computing device (see Table 1).

Importantly, the deactivation rate of the off state is small yet must be nonzero for thermodynamic consistency (17). We also note that for the case where proteins are activated through phosphorylaltion, includes nonspecific phosphorylation arising from other kinases as well as contributions from the reverse reactions of the phosphotases. The inactivation rate sets the scale for the effective measurement time because it is the rate at which information encoded in downstream proteins is lost due to inactivation. In order to compute external concentrations accurately, the measurement time must be much longer than the typical switching times between receptor states, . We show below that this simple network in fact implements a noisy version of the original Berg–Purcell calculation. Our explicit construction allows us to study the relationship between information and power consumption in this network.

The paper is organized as follows. In the first section, we compute the steady-state behavior of the system, first within the linear-noise approximation and then through an exact solution of the corresponding master equation. In the next two sections, we quantify the efficacy of chemosensation through the variance in estimated concentration (using the linear-noise approximation) and calculate the power consumption required to maintain the network in steady state (using the exact solution). Finally, we show there exists a tradeoff between these two quantities and discuss implications for the design of cellular sensing systems. We begin each section with a general discussion that summarizes the major ideas, results, and interpretations followed by a more technical and mathematical discussion.

## Characterization of the Network’s Steady-State Properties

In this section, we derive the steady-state properties of the cellular network considered above. The network translates the external ligand concentration into an internal concentration of activated proteins. This mapping between external ligand concentration and downstream protein number is probabilistic due to the stochasticity inherent in biochemical networks. For this reason, we can characterize the output of the network by a probability distribution of activated proteins, *p*(*n*). The shape of the distribution *p*(*n*) depends explicitly on the kinetic parameters (*k*_{1}, , ) (see Fig. 2). In the following sections, we focus on the fast-switching regime where receptors switch between the on and off states quickly as compared to the deactivation rate, and as we will see, *p*(*n*) is unimodal (see Fig 2*B*).

In this regime, two important characteristics of the distribution *p*(*n*) are the mean, protein number, , and variance . The mean protein number is the cell’s best estimate of the external ligand concentration. On the other hand, the variance characterizes the cell’s uncertainty about external ligand concentrations due to stochasticity in the underlying biochemical network. Though it is possible to decrease the variance by increasing the mean protein number, it will be always be nonzero due to fluctuations in the state of the receptors (12).

We begin by first deriving the steady-state mean and variance of the number of activated proteins, *n*, within the linear-noise approximation. Afterwards, we will study the full probability distribution. The deterministic dynamics of the biochemical network in Fig. 1 is captured by simple rate-equations for the time dependence of the receptor state probabilities and the mean number of activated proteins. We can augment these equations to account for stochastic fluctuations within the linear-noise approximation by adding appropriate Langevin noise terms. In the remainder of this paper, we assume for simplicity that proteins are abundant and ignore saturation effects. The dynamics of the circuit is therefore described by a pair of Langevin equations for the probabilities *p*_{on} and *p*_{off} (i.e. 1 - *p*_{on}) for the receptor to be in the on or off states, respectively, together with the number of activated proteins *n*, [1][2]The variance of the Langevin terms is given by the Poisson noise in each of the reactions: [3]where δ(*t* - *t*^{′}) denotes the Dirac-delta function and the overbar denotes the mean steady-state value of the respective quantity (18, 19).

At steady-state, we can calculate the mean probability and number of proteins by setting the time derivative in Eq. **2** equal to zero while ignoring noise terms, yielding [4]and [5]where we have defined the dimensionless parameters with *j* = {2,4} and *s* = {on,off}. For the biologically realistic case , the mean number of proteins is simply proportional to the kinase activity in the on state times the probability of being in the on state, , as expected. One can further calculate the variance in protein numbers (*SI Text*) [6]The first term on the right-hand side of the equation results from Poisson noise in the synthesis and degradation of activated protein, whereas the second term is due to stochastic fluctuations in the state of the receptors.

In addition to the mean and variance, we will need the full steady-state probability distribution for *n* to calculate the power consumption of the network. The steady-state distribution can be calculated from the master equation for the probability, *p*_{s}(*n*), of there being *n* active proteins with the receptor in a state *s*: [7] where (on) when *s* = on (off). This equation is similar to those found in (20, 21) and the steady-state distribution can be solved via a generating function approach (see *SI Text*).

Depending on the parameters, the steady-state distributions can have two qualitatively distinct behaviors as shown in Fig. 2. In the “slow switching” regime with and , receptors switch at rates much slower than the protein deactivation rate *k*_{1}. The result is a bimodal distribution of activated proteins that can be intuitively understood to arise from the superposition of the probability distributions of activated proteins when the receptor is the on and off states. Similar behavior has been found in the steady-state expression of a self-regulation of a gene (22). As approaches , the distributions in the two states merge and the overall probability distribution becomes unimodal. On the other hand, in the “fast switching” regime, characterized by , the distribution of activated proteins is always unimodal. In this limit, the measurement time, , is much longer than the average time a receptor remains in the on or off state, and the biochemical network “time-averages” out the stochastic fluctuations in receptor states. In what follows, we restrict our considerations to this latter regime.

## Quantification of Learning

The biochemical circuit in Fig. 1 “computes” the external concentration of a chemical ligand. As emphasized by Berg and Purcell in their seminal paper (12), the chief obstacle in determining external concentration is the stochastic fluctuations in the state of the ligand-binding receptors. Berg and Purcell argued that a reasonable measure of how much cells learn is the uncertainty cells have about external concentration as measured by the variance of the estimated concentration, (δ*c*)^{2}. (δ*c*)^{2} measures the uncertainty about the external ligand concentration based on the probability distribution of downstream activated proteins. Using standard arguments, we show below that this uncertainty is directly related to the variance, (δ*n*)^{2}, of the corresponding protein probability distribution. This framework allows us to use the results of the last section to quantify how much cells learn about external ligand concentrations as a function of kinetic parameters. The results are plotted in Fig. 2*C*.

To compute uncertainty, Berg and Purcell assumed that the cell computes the average receptor occupancy by time-averaging over a measurement time *T*. They showed (12) that [8]where , is independent of *c*, and *N*_{b} is the number of binding events during the time *T*. It was later shown that cells could compute concentration more accurately by implementing MLE with (14, 23) [9]The factor of two decrease in uncertainty derives from the fact that MLE ignores noise due to unbinding of ligands from the cell. We note that this "forgetting" of ligand unbinding, in turn, requires the receptors themselves to be out of equilibrium, further contributing to the system’s energy consumption. We do not, however, consider the additional energetic costs of nonequilibrium receptors here.

To quantify learning in our biochemical circuit, we follow Berg and Purcell and estimate the fluctuations in (δ*c*)^{2} as [10]with . Substituting and and computing the derivative using Eq. **5** gives [11]Substituting Eqs. **5** and **6** into Eq. **10** yields [12]Similar to the linear-noise calculation, the first term on the right-hand side arises from the Poisson fluctuations in activated protein number, while the second term results from the stochastic fluctuations in the state of receptors. Fig. 2 shows the uncertainty, (δ*c*)^{2}/*c*^{2}, as a function of the degradation rate of activated protein, *k*_{1}, when and and .

By identifying the degradation rate with the inverse measurement time, *k*_{1} = 2*T*^{-1}, we can also compare the results with Berg–Purcell. The factor of two is due to the slight difference in how the variance of the average receptor occupancy is calculated for a biochemical network when compared to the original Berg–Purcell calculation (23). As shown in Fig. 2, when is increased, the Poisson noise in protein production is suppressed and the performance of the cellular network approaches that due to Berg-Purcell. To make the connection with Berg–Purcell more explicit, it is helpful to rewrite Eq. **12** in terms of the average number of binding events, *N*_{b}, during the averaging time, *T*: [13]This form allows us to see the approach to the Berg–Purcell limit. When the measurement time is much longer than the timescale of fluctuations in receptor activity, i.e., (or equivalently ) and the average number of activated proteins is large, , the expression above reduces to (δ*c*)^{2}/*c*^{2} ≈ 2/*N*_{b} in agreement with Eq. **8**.

## Power Consumption and Entropy Production in Steady State

We now compute the energy consumed by the circuit in Fig. 1 as a function of the kinetic parameters. To do so, we exploit the fact that dynamics of the circuit can be formulated as a nonequilbrium Markov process (see Fig. 3). A nonequilibrium steady-state (NESS) necessarily implies the breaking of detailed balance in the underlying Markovian dynamics and, therefore, possesses a nonzero entropy production rate. The entropy production rate is precisely the amount of power consumed by the biochemical circuit in maintaining the nonequilibrium steady state. Thus, by calculating the entropy production rate as function of kinetic parameters, we can calculate the power consumed by the biochemical network implementing the computation. To calculate entropy production, we utilize the full steady-state probability distribution and standard formulas from the theory on nonequilibrium Markov processes.

Consider a general Markov process with states labeled by σ and transition probability from σ to σ^{′} given by *k*(σ,σ^{′}). Defining the steady-state probability of being in state σ by *P*_{σ}, the entropy production rate, *EP*, for a NESS is given by (24) [14]For our problem this general formula reduces to [15]where we again use (see *SI Text*). In deriving this formula, we assumed that the receptors were in thermal equilibrium and obeyed detailed balance. Using explicit expressions for the steady-state distributions, *p*_{s}(*n*), we can calculate the energy consumption of the network as a function of kinetic parameters (see *SI Text*). The physical content of this expression is summarized in Fig. 3. The expression states that any nonzero cyclic flux must necessarily produce entropy. Otherwise, one would have a chemical version of a perpetual motion machine. Figs. 3 and 4 show the power consumption as a function of and *k*_{1}. Notice that the power consumption tends to zero as both these parameters go to zero. Note that we cannot, however, set *k*_{1} = 0 identically because there then no longer exists a steady-state distribution.

## Energetics, Information, and Landauer’s Principle

We now highlight the fundamental connection between the energy consumed by the network and the information the network acquires about the environment and briefly discuss its relation to Landauer’s principle. First, note that learning information about the environment requires energy consumption by the network. This relationship can be seen in Fig. 3, which shows that as Δ*k*_{2} → 0, the uncertainty about the concentration tends to infinity. Moreover, we show in the *SI Text* that the entropy production, Eq. **15**, is zero if and only if Δ*k*_{2} = 0. In conjunction with Eq. **12**, which diverges as Δ*k*_{2} → 0, these observations imply that learning requires consuming energy. Mathematically, in the limit where Δ*k*_{2} = 0, the dynamics of the Markov process in Fig. 3 become “one-dimensional” instead of a two-legged ladder, and the dynamics obeys detailed balance. Physically, in this limit the number of downstream proteins becomes insensitive to external ligand concentrations because all information about concentration is contained in the relative probabilities of being in the on or off state.

Second, as shown in Fig. 4, the power consumption of the circuit tends to zero as *k*_{1} → 0. This result is consistent with, and a manifestation of, Landauer’s principle: Entropy production stems from erasing memory in a computing device. The number of activated proteins serves the function of a memory of ligand concentration, which is erased at the dephosphorylation rate *k*_{1}. Thus, as the erasure rate of the memory tends to zero, the device consumes less energy per unit time, as expected. Yet despite the fact that the power consumption tends to zero as *k*_{1} decreases, the total energy consumed per measurement, namely the power times the measurement time, , still increases (see Fig. 4). Thus, learning more requires consuming more total energy despite the fact that power consumption is decreasing. In effect, one is approaching the reversible computing limit where memory is erased adiabatically. Again note, however, that when erasure is performed infinitely slowly, *k*_{1} = 0, the system no longer has an NESS and our formalism does not apply.

Finally, we note that one of the important open problems in our understanding of chemosensation is a full characterization of the constraints placed on the measurement time *T*. In principle, cells can always learn more by measuring the environment for longer periods of time. However, in most biological systems, these measurement times are observed to be quite short. There are a number of constraints that can limit this measurement time, including rotational diffusion (12) (in the case of swimming bacteria) and the restrictions placed on motility. Here, we highlight another restriction that may be important in resource-starved environments: Sensing external concentration necessarily requires cells to consume energy.

## Discussion and Conclusion

Cells often perform computations using elaborate biochemical networks that respond to environmental cues. One of the most common simple networks found in bacteria are two-component networks where a receptor phosphorylates a downstream response regulator (16). In this work, we have shown that these simple two-component networks can implement a noisy version of the Berg–Purcell strategy to compute the concentration of external ligands. Furthermore, by mapping the dynamics of the biochemical network to nonequilibrium steady-states in Markov processes, we explicitly derived expressions for the power consumed by the network and showed that learning requires energy consumption. Taken together, these calculations suggest that, much like man-made and neural computing (3, 9⇓–11), energetic considerations may place important constraints on the design of biochemical networks that implement cellular computations. They also suggest a fundamental tradeoff between the efficiency of cellular computing and the requisite energy consumption.

Bacterial cells such as *Bacillus subtilis* can sporulate during times of environmental stress and remain metabolically dormant for many years. Although sporulation is relatively well understood, the reverse process of germination is much more difficult to study. One current model for how a spore knows when to germinate in response to external cues involves integrating the signal and triggering commitment when an accumulation threshold is reached (25, 26). Such a scheme corresponds to the limit of vanishingly small *k*_{1} in our model, so that power consumption is minimized at the expense of retaining the entire integrated signal. Our results indicate that this behavior may be due to the extreme energetic constraints imposed on a metabolically dormant spore, rather than an evolutionarily optimized strategy.

An important insight of this work is that even a simple Berg–Purcell strategy for sensing external concentrations requires the consumption of energy. It is likely that more complicated strategies that increase how much cells learn, such as maximum likelihood, require additional energetic inputs. For example, it was argued in (23) that MLE can be implemented by a network similar to the perfect adaptation network where bursts are produced in response to binding events. These bursts break detailed balance and therefore require energy consumption. It will be interesting to investigate further how the tradeoff between learning and energy consumption manifests itself in the design of computational strategies employed by cells.

In this work, we restricted ourselves to the simple case where cells calculate the steady-state concentration of an external signal. In the future, it will be useful to generalize the analysis performed here to other computations such as responding to temporal ramps (23) and spatial gradients (27, 28). It will also be interesting to understand how to generalize the considerations here to arbitrary biochemical networks. An important restriction on our work is that we reduced our considerations to nonequilibrium steady states. It will be interesting to ask how to generalize the work here to biochemical networks with a strong temporal component.

## Acknowledgments

P.M. and D.J.S. would like to thank the Aspen Center for Physics, where this work was initiated. We are especially grateful to Thierry Mora for clarifying the relationship between the rate *k*_{1} and the average integration time *T*. This work was partially supported by National Institute of Health Grants K25GM086909 (to P.M.). DS was partially supported by Defense Advanced Research Projects Agency grant HR0011-05-1-0057 and National Science Foundation grant PHY-0957573.

## Footnotes

- ↵
^{1}To whom correspondence should be addressed. E-mail: pankajm{at}bu.edu.

Author contributions: P.M. and D.J.S. designed research, performed research, contributed new reagents/analytic tools, analyzed data, and wrote the paper.

The authors declare no conflict of interest.

This article is a PNAS Direct Submission.

This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1207814109/-/DCSupplemental.

## References

- ↵
- ↵
- ↵
- ↵
- ↵
- ↵
- ↵
- ↵
- Tu Y

- ↵
- ↵
- ↵
- ↵
- ↵
- Bialek W,
- Setayeshgar S

- ↵
- ↵
- ↵
- ↵
- Beard D,
- Qian H

- ↵
- ↵
- ↵
- ↵
- ↵
- Hornos JE,
- et al.

- ↵
- ↵
- ↵
- Yi X,
- Liu J,
- Faeder J,
- Setlow P

- ↵
- ↵
- Endres R,
- Wingreen N

- ↵

## Citation Manager Formats

### More Articles of This Classification

### Biological Sciences

### Biophysics and Computational Biology

### Physical Sciences

### Related Content

- No related articles found.