Theory for the optimal detection of time-varying signals in cellular sensing systems
Abstract
Living cells often need to measure chemical concentrations that vary in time, yet how accurately they can do so is poorly understood. Here, we present a theory that fully specifies, without any adjustable parameters, the optimal design of a canonical sensing system in terms of two elementary design principles: (1) there exists an optimal integration time, which is determined by the input statistics and the number of receptors; and (2) in the optimally designed system, the number of independent concentration measurements as set by the number of receptors and the optimal integration time equals the number of readout molecules that store these measurements and equals the work to store these measurements reliably; no resource is then in excess and hence wasted. Applying our theory to the Escherichia coli chemotaxis system indicates that its integration time is not only optimal for sensing shallow gradients but also necessary to enable navigation in these gradients.
Introduction
Living cells continually have to respond and adapt to changes in their environment. They often do so on a timescale that is comparable to that of the environmental variations. Examples are cells that during their development differentiate in response to time-varying morphogen gradients (Durrieu et al., 2018) or cells that navigate through their environment (Tostevin and ten Wolde, 2009; Sartori and Tu, 2011; Long et al., 2016). These cells shape, via their movement, the statistics of the input signal, such that the timescale of the input fluctuations becomes comparable to that of the response. In all these cases, it is important to understand how accurately the cell can estimate chemical concentrations that vary in time.
Cells measure chemical concentrations via receptors on their surface. These measurements are inevitably corrupted by the stochastic arrival of the ligand molecules by diffusion and by the stochastic binding of the ligand to the receptor. Wiener and Kolmogorov (Extrapolation, 1950; Kolmogorov, 1992) and Kalman, 1960 have developed theories for the optimal strategy to estimate signals in the presence of noise. Their filtering theories have been employed widely in engineering, and in recent years they have also been applied to cell signaling. They have been used to show that time integration can improve the sensing of time-varying signals by reducing receptor noise, although it cannot remove this input noise completely because of signal distortion (Andrews et al., 2006; Hinczewski and Thirumalai, 2014; Becker et al., 2015). It has been shown that circadian systems can adapt their response to the statistics of the input signal, as predicted by Kalman filtering theory (Husain et al., 2019). Moreover, Wiener–Kolmogorov filtering theory has been employed to derive the optimal topology of the cellular network depending on the statistics of the input signal (Becker et al., 2015). Negative feedback and incoherent feedforward, which are common motifs in cell signaling (Alon, 2007), make it possible to predict future signal values via signal extrapolation, which is useful when the past signal contains information about the future in addition to the current signal (Becker et al., 2015).
The precision of sensing depends not only on the topology of the cellular sensing network but also on the resources required to build and operate it. Receptors and time are needed to take the concentration measurements (Berg and Purcell, 1977), downstream molecules are necessary to store the ligand-binding states of the receptor in the past, and energy is required to store these states reliably (Govern and Ten Wolde, 2014a). Many studies have addressed the question how receptors and time limit the precision of sensing static concentrations that do not vary on the timescale of cellular response (Berg and Purcell, 1977; Bialek and Setayeshgar, 2005; Wang et al., 2007; Rappel and Levine, 2008; Endres and Wingreen, 2009; Hu et al., 2010; Mora and Wingreen, 2010; Govern and Ten Wolde, 2012; Mehta and Schwab, 2012; Govern and Ten Wolde, 2014a; Govern and Ten Wolde, 2014b; Kaizu et al., 2014; Ten Wolde et al., 2016; Mugler et al., 2016; Fancher and Mugler, 2017). In addition, progress has been made in understanding how the number of readout molecules and energy set the precision of sensing static signals (Mehta and Schwab, 2012; Govern and Ten Wolde, 2014a; Govern and Ten Wolde, 2014b). Yet, what the resource requirements for sensing time-varying signals are is a wide open question. In particular, it is not known how the number of receptor and readout molecules, time, and power required to maintain a desired sensing precision depend on the strength and the timescale of the input fluctuations.
In this article, we present a theory for the optimal design of cellular sensing systems as set by resource constraints and the dynamics of the input signal. The theory applies to one of the most common motifs in cell signaling, a receptor that drives a push–pull network, which consists of a cycle of protein activation and deactivation (Goldbeter and Koshland, 1981, see Figure 1). These systems are omnipresent in prokaryotic and eukaryotic cells (Alon, 2007). Examples are GTPase cycles, as in the Ras system, phosphorylation cycles, as in MAPK cascades, and two-component systems like the chemotaxis system of Escherichia coli. Push–pull networks constitute a simple exponential filter (Hinczewski and Thirumalai, 2014; Becker et al., 2015), in which the current output depends on the current and past input (with past input values contributing to the output with a weight that decays exponentially with time back into the past). Wiener–Kolmogorov filtering theory (Extrapolation, 1950; Kolmogorov, 1992) shows that these networks are optimal for estimating signals that are memoryless (Becker et al., 2015), meaning that the past input does not contain information that is not already present in the current input. These networks are useful because they act as low-pass filters, removing the high-frequency receptor–ligand-binding noise (Andrews et al., 2006; Hinczewski and Thirumalai, 2014; Becker et al., 2015). Push–pull networks thus enable the cell to employ the mechanism of time integration, in which the cell infers the concentration not from the instantaneous number of ligand-bound receptors, but rather from the average receptor occupancy over an integration time (Berg and Purcell, 1977). Our theory gives a unified description in terms of all the cellular resources, protein copies, time, and energy, that are necessary to implement this mechanism of time integration. It does not address the sensing strategy of maximum-likelihood estimation (Endres and Wingreen, 2009; Mora and Wingreen, 2010; Lang et al., 2014; Hartich and Seifert, 2016; Ten Wolde et al., 2016) or Bayesian filtering (Mora and Nemenman, 2019).
While filtering theories are powerful tools for predicting the optimal topology and response dynamics of the cellular sensing network (Andrews et al., 2006; Hinczewski and Thirumalai, 2014; Becker et al., 2015), they do not naturally reveal the resource requirements for sensing. Our theory therefore employs the sampling framework of Govern and Ten Wolde, 2014a and extends it here to time-varying signals. This framework is based on the observation that the cell estimates the current ligand concentration not from the current number of active readout molecules directly, but rather via the receptor: the cell uses its push–pull network to estimate the receptor occupancy from which the ligand concentration is then inferred (see Figure 2). To elucidate the resource requirements for time integration, the push–pull network is viewed as a device that employs the mechanism of time integration by discretely sampling, rather than continuously integrating, the state of the receptor via collisions of the readout molecules with the receptor proteins (see Figure 2). During each collision, the ligand-binding state of the receptor protein is copied into the activation state of the readout molecule (Ouldridge et al., 2017). The readout molecules thus constitute samples of the receptor state, and the fraction of active readout molecules provides an estimate of the average receptor occupancy. The readout activation states have, however, a finite lifetime, which means that this is an estimate of the (running) average receptor occupancy over this lifetime, which indeed sets the receptor integration time . The cell can estimate the current ligand concentration L from this estimate of the average receptor occupancy over the past integration time because there is a unique one-to-one mapping between and L. This mapping is the dynamic input–output relation and differs from the conventional static input–output relations used to describe the sensing of static concentrations that do not vary on the timescale of the response (Berg and Purcell, 1977; Bialek and Setayeshgar, 2005; Kaizu et al., 2014; Ten Wolde et al., 2016) in that it depends not only on the response time of the system but also on the dynamics of the input signal.
Our theory reveals that the sensing error can be decomposed into two terms, which each depend on collective variables that reveal the resource requirements for sensing. One term, the sampling error, describes the sensing error that arises from the finite accuracy by which the receptor occupancy is estimated. This error depends on the number of receptor samples, as set by the number of receptors, readout molecules, and the integration time; their independence, as given by the receptor-sampling interval and the timescale of the receptor–ligand-binding noise; and their reliability, as determined by how much the system is driven out of thermodynamic equilibrium via fuel turnover. The other term is the dynamical error and is determined by how much the concentration in the past integration time reflects the current concentration that the cell aims to estimate; it depends on the integration time and timescale of the input fluctuations.
Our theory gives a comprehensive view on the optimal design of a cellular sensing system. Firstly, it reveals that the resource allocation principle of Govern and Ten Wolde, 2014a can be generalized to time-varying signals. There exist three fundamental resource classes – receptors and their integration time, readout molecules, and power and integration time – which each fundamentally limit the accuracy of sensing; and, in an optimally designed system, each resource class is equally limiting so that none of them is in excess and thus wasted. However, in contrast to sensing static signals, time cannot be freely traded against the number of receptors and the power to achieve a desired sensing precision: there exists an optimal integration time that maximizes the sensing precision, which arises as a trade-off between the sampling error and dynamical error. Together with the resource allocation principle, it completely specifies, without any adjustable parameters, the optimal design of the system in terms of its resources protein copies, time, and energy.
Our theory also makes a number of specific predictions. The optimal integration time decreases as the number of receptors is increased because this allows for more instantaneous measurements. Moreover, the allocation principle reveals that when the input varies more rapidly both the number of receptors and the power must increase to maintain a desired sensing precision, while the number of readout molecules does not.
Finally, we apply our theory to the chemotaxis system of E. coli. This bacterium searches for food via a run-and-tumble strategy (Berg and Brown, 1972), yielding a fluctuating input signal. In small gradients, the timescale of these input fluctuations is set by the typical run time of the bacterium, which is on the order of a few seconds (Berg and Brown, 1972; Taute et al., 2015), while the strength of these fluctuations is determined by the steepness of the gradient. Interestingly, experiments have revealed that E. coli can sense extremely shallow gradients, with a length scale of approximately 104µm (Shimizu et al., 2010), raising the question how accurately E. coli can measure the concentration and whether this is sufficient to determine whether during a run it has changed, even in these shallow gradients. To measure the concentration, the chemotaxis system employs a push–pull network to filter out the high-frequency receptor–ligand-binding noise (Sartori and Tu, 2011). Applying our theory to this system predicts that the measured integration time, on the order of 100 ms (Sourjik and Berg, 2002), is not only sufficient to enable navigation in these shallow gradients but also necessary. This suggests that this system has evolved to optimally sense shallow concentration gradients.
Results
Theory: model
We consider a single cell that needs to sense a time-varying ligand concentration (see Figure 1a). The ligand concentration dynamics is modeled as a stationary memoryless, or Markovian, signal specified by the mean (total) ligand concentration , the variance , and the correlation time , which determines the timescale on which input fluctuations decay. It obeys Gaussian statistics (Tostevin and ten Wolde, 2010).
The concentration is measured via receptor proteins on the cell surface, which independently bind the ligand (Ten Wolde et al., 2016), . The correlation time of the receptor state, which is the timescale on which fluctuations in the number of ligand-bound receptors regresses to the mean, is given by (Berg and Purcell, 1977; Bialek and Setayeshgar, 2005; Kaizu et al., 2014; Ten Wolde et al., 2016). It determines the timescale on which independent concentration measurements can be made.
The ligand-binding state of the receptor is read out via a push–pull network (Goldbeter and Koshland, 1981). The most common scheme is phosphorylation fueled by the hydrolysis of adenosine triphosphate (ATP) (see Figure 1b). The receptor, or an enzyme associated with it such as CheA in E. coli, catalyzes the modification of the readout, . The active readout proteins can decay spontaneously or be deactivated by an enzyme, such as CheZ in E. coli, . Inside the living cell the system is maintained in a non-equilibrium steady state by keeping the concentrations of ATP, adenosine diphosphate (ADP), and inorganic phosphate (Pi) constant. We absorb their concentrations and the activities of the kinase and, if applicable, phosphatase in the (de)phosphorylation rates, coarse-graining the (de)modification reactions into instantaneous second-order reactions: , . This system has a relaxation time (Govern and Ten Wolde, 2014a), which describes how fast fluctuations in relax. It determines how long can carry information on the ligand-binding state of the receptor; thus sets the integration time of the receptor state.
Theory: inferring concentration from receptor occupancy
The central idea of our theory is illustrated in Figure 2a: the cell employs the push–pull network to estimate the average receptor occupancy over the past integration time . It then uses this estimate to infer the current concentration L via the dynamic input–output relation , which provides a one-to-one mapping between and L.
Dynamic input–output relation
The mapping is the dynamic input–output relation. It gives the average receptor occupancy over the past integration time , given that the current value of the input signal is (see Figure 2a). Here, the average is not only over the noise in receptor–ligand binding and readout activation (Figure 2b) but also over the subensemble of past input trajectories that each end at the same current concentration L (Figure 2c; Tostevin and ten Wolde, 2010; Hilfinger and Paulsson, 2011; Bowsher et al., 2013). In contrast to the conventional static input–output relation , which gives the average receptor occupancy p for a steady-state ligand concentration that does not vary in time, the dynamic input–output relation takes into account the dynamics of the input and the finite response time of the system. It depends on all timescales in the problem: the timescale of the input, , the receptor–ligand correlation time , and the integration time . Only when does the dynamic input–output become equal to the static input–output relation .
Sensing error
Linearizing the dynamic input–output relation around the mean ligand concentration (see Figure 2a) and using the rules of error propagation, the expected error in the concentration estimate is
Here, is the variance in the estimate of the average receptor occupancy over the past , given that the current input signal is L (see Figure 2a). The quantity is the dynamic gain, which is the slope of the dynamic input–output relation ; it determines how much an error in the estimate of propagates to that in L. Equation 1 generalizes the expression for the error in sensing static concentrations (Berg and Purcell, 1977; Bialek and Setayeshgar, 2005; Wang et al., 2007; Mehta and Schwab, 2012; Kaizu et al., 2014; Govern and Ten Wolde, 2014a; Ten Wolde et al., 2016) to that of time-varying concentrations.
Signal-to-noise ratio
Together with the distribution of input states, the sensing error determines how many distinct signal values the cell can resolve. The latter is quantified by the signal-to-noise ratio (SNR), which is defined as
Here, is the variance of the ligand concentration ; because the system is stationary and time invariant, we can omit the argument in and write . The variance is a measure for the total number of input states, such that the SNR gives the number of distinct ligand concentrations the cell can measure. Using Equation 1, it is given by
The SNR also yields the mutual information between the input L and output (Tostevin and ten Wolde, 2010).
Readout system samples receptor state
Receptor time averaging is typically conceived as a scheme in which the receptor state is averaged via the mathematical operation of an integral: . Yet, readout proteins are discrete components that interact with the receptor in a discrete and stochastic fashion. To derive the dynamic gain and error in estimating , (Equation 3), we therefore view the push–pull network as a device that discretely samples the receptor state (see Figure 2b; Govern and Ten Wolde, 2014a). The principle is that cells employ the activation reaction to store the state of the receptor in stable chemical modification states of the readout molecules. Readout molecules that collide with a ligand-bound receptor are modified, while those that collide with an unbound receptor are not (Figure 2b). The readout molecules serve as samples of the receptor at the time they were created, and collectively they encode the history of the receptor: the fraction of samples that correspond to ligand-bound receptors is the cell’s estimate for . Indeed, this is the discrete and stochastic implementation of the mechanism of time integration. The effective number of independent samples depends not only on the creation of samples, , but also on their decay and accuracy. Samples decay via the deactivation reaction , which means that they only provide information on the receptor occupancy over the past . In addition, both the activation and the deactivation reaction can happen in their microscopic reverse direction, which corrupts the coding, that is, the mapping between the ligand-binding states of the receptor proteins and the activation states of the readout molecules. Energy is needed to break time reversibility and protect the coding. Furthermore, for time-varying signals, we also need to recognize that the samples correspond to the ligand concentration over the past integration time , which will in general differ from the current concentration L that the cell aims to estimate (see Figure 2c). While a finite is necessary for time integration, it will, as we show below, also lead to a systematic error in the estimate of the concentration that the cell cannot reduce by taking more receptor samples.
This analysis reveals that the dynamic gain is (see Appendix 1)
Only when is the average ligand concentration over the ensemble of trajectories ending at equal to the current concentration (Figure 2c) and does become equal to its maximal value, the static gain , where p is the average receptor occupancy averaged over all values of . The analysis also reveals that the error in can be written as (see Appendix 1, Equation 29)
where is a statistical error due to the stochastic sampling of the receptor and is a systematic error arising from the dynamics of the input, as elucidated in Figure 2b, c.
Central result
To know how the error in the estimate of propagates to the error in the estimate of the current ligand concentration, we divide by the dynamic gain given by Equation 4 (see Equation 1). For the full system, the reversible push–pull network, this yields via Equation 3 the central result of our article, the SNR in terms of the total number of receptor samples, their independence, their accuracy, and the timescale on which they are generated:
This expression shows that the sensing error can be decomposed into two distinct contributions, which each have a clear interpretation: the sampling error, arising from the stochasticity in the sampling of the receptor state, and the dynamical error, arising from the dynamics of the input.
When the timescale of the ligand fluctuations is much longer than the receptor correlation time and the integration time , , the dynamical error reduces to zero and only the sampling error remains. Here, is the total number of effective samples and is the number of these that are independent (Govern and Ten Wolde, 2014a). For the full system, they are given by
The quantity is the net flux of x around the cycle of activation and deactivation, with the total number of receptor proteins and and the average number of inactive and active readout molecules, respectively. It equals the rate at which x is modified by the ligand-bound receptor; the quantity is thus the sampling rate of the receptor, be it ligand bound or not. Multiplied with the relaxation rate , it yields the total number of receptor samples obtained during . However, not all these samples are reliable. The effective number of samples is , where quantifies the quality of the sample. Here, is the inverse temperature, and are the free-energy drops over the activation and deactivation reaction, respectively, with the total drop, determined by the fuel turnover (see Figure 1b). If the system is in thermodynamic equilibrium, , and the system cannot sense because the ligand-binding state of the receptor is equally likely to be copied into the correct modification state of the readout as into the incorrect one. In contrast, if the system is strongly driven out of equilibrium and , then, during each receptor–readout interaction, the receptor state is always copied into the correct activation state of the readout; the sample quality parameter q thus approaches unity and . Yet, even when all samples are reliable, they may contain redundant information on the receptor state. The factor is the fraction of the samples that are independent. It reaches unity when the receptor sampling interval becomes larger than the receptor correlation time .
When the number of samples becomes very large, , the sampling error reduces to zero. However, the sensing error still contains a second contribution, which, following Bowsher et al., 2013, we call the dynamical error. This contribution only depends on timescales. It arises from the fact that the samples encode the receptor history and hence the ligand concentration over the past , which will, in general, deviate from the quantity that the cell aims to predict – the current concentration L. This contribution yields a systematic error, which cannot be eliminated by increasing the number of receptor samples, their independence, or their accuracy. It can only be reduced to zero by making the integration time much smaller than the ligand timescale (assuming is typically much smaller than ). Only in this regime will the ligand concentration in the past be similar to the current concentration and can the latter be reliably inferred from the receptor occupancy, provided the latter has been estimated accurately by taking enough samples.
Importantly, the dynamics of the input signal not only affects the sensing precision via the dynamical error but also via the sampling error. This effect is contained in the prefactor of the sampling error, , which has its origin in the dynamic gain (Equation 4). It determines how the sampling error in the estimate of propagates to the error in the estimate of L (see Equation 3). Only when can the readout system closely track the input signal and does reach its maximal value, the static gain , thus minimizing the error propagation from to L.
Fundamental resources
We can use Equation 6 to identify the fundamental resources for cell sensing (Govern and Ten Wolde, 2014a) and derive Pareto fronts that quantify the trade-offs between the maximal sensing precision and these resources. A fundamental resource is a (collective) variable that, when fixed to a constant, puts a non-zero lower bound on , no matter how the other variables are varied. It is thus mathematically defined as To find these collective variables, we numerically or analytically minimized , constraining (combinations of) variables yet optimizing over the other variables. This reveals that the SNR is bounded by (see Appendix 2)
where
Equations 8 and 9 show that the fundamental resources are the number of receptors , the integration time , the number of readouts , and the power .
Figure 3a, b illustrates that are indeed fundamental: the sensing precision is bounded by the limiting resource and cannot be enhanced by increasing another resource. Panel (a) shows that when is small, the maximum mutual information cannot be increased by raising : no matter how many receptors the system has, the sensing precision is limited by the pool of readout molecules and only increasing this pool can raise . Yet, when is large, becomes independent of . In this regime, the number of receptors limits the number of independent concentration measurements and only increasing can raise . Similarly, panel (b) shows that when the power is limiting, cannot be increased by but only by increasing . Clearly, the resources receptors, readout molecules, and energy cannot compensate each other: the sensing precision is bounded by the limiting resource.
Importantly, while for sensing static concentrations the products and are fundamental (Govern and Ten Wolde, 2014a), for time-varying signals , , and separately limit sensing. Consequently, neither receptors nor power can be traded freely against time to reach a desired precision, as is possible for static signals. In line with the predictions of signal filtering theories (Extrapolation, 1950; Kolmogorov, 1992; Kalman, 1960), there exists an optimal integration time that maximizes the sensing precision (Andrews et al., 2006; Hinczewski and Thirumalai, 2014; Becker et al., 2015; Monti et al., 2018b; Mora and Nemenman, 2019). Interestingly, its value depends on which of the resources , , and is limiting (Figure 3c–f). We now discuss these three regimes in turn.
Receptors
Berg and Purcell, 1977 pointed out that cells can reduce the sensing error by either increasing the number of receptors or taking more measurements per receptor via the mechanism of time integration. However, Equation 8 reveals that for sensing time-varying signals time integration can never eliminate the sensing error completely, as predicted also by filtering theories (Extrapolation, 1950; Kolmogorov, 1992; Kalman, 1960). Equation 8 shows that in the Berg–Purcell regime, where receptors and their integration time are limiting and , the sensing precision does not depend on , as for static signals (Govern and Ten Wolde, 2014a), but on and separately, such that an optimal integration time emerges that maximizes the sensing precision (see Figure 3c). Increasing improves the mechanism of time integration by increasing the number of independent samples per receptor, , thus reducing the sampling error (Equation 6). However, increasing raises the dynamical error. Moreover, it lowers the dynamical gain , which increases the propagation of the error in the estimate of the receptor occupancy to that of the ligand concentration. The optimal integration time arises as a trade-off between these three factors.
Figure 3c also shows that the optimal integration time decreases with the number of receptors . The total number of independent concentration measurements is the number of independent measurements per receptor, , times the number of receptors, . As increases, less measurements per receptor have to be taken to remove the receptor–ligand-binding noise, explaining why decreases as increases – time integration becomes less important.
Interestingly, depends non-monotonically on the receptor–ligand correlation time (Figure 3d). When increases at fixed , the receptor samples become more correlated. To keep the mechanism of time integration effective, must increase with . However, to avoid too strong signal distortion the cell compromises on time integration by decreasing the ratio (see inset). When becomes too large, the benefit of time integration no longer pays off the cost of signal distortion. Now not only the ratio decreases but also itself. The sensing system switches to a different strategy: it no longer employs time integration but becomes an instantaneous sensor.
Readout molecules
To implement time integration, the cell needs to store the receptor states in the readout molecules. When the number of readout molecules is limiting, the sensing precision is given by Equation 8 with . This bound is saturated when . This is in marked contrast to the non-zero optimal integration in the Berg–Purcell regime (see Figure 3c).
To elucidate the non-trivial behavior of , Figure 3e shows as a function of . When is smaller than , the average number of samples per receptor is less than unity. In this regime, the system cannot time integrate the receptor, and to minimize signal distortion . Yet, when is increased, the likelihood that two or more readout molecules provide a sample of the same receptor molecule rises, and time averaging becomes possible. Yet to obtain receptor samples that are independent, the integration time must be increased to make the sampling interval larger than the receptor correlation time . As and hence the total number of samples are increased further, the number of samples that are independent, , only continues to rise when increases with further. However, while this reduces the sampling error, it also increases the dynamical error. When the decrease in the sampling error no longer outweighs the increase in the dynamical error, and the mutual information no longer change with (see Figure 3a). The system has entered the Berg–Purcell regime in which and the mutual information are given by the optimization of Equation 8 with (gray dashed line). In this regime, increasing merely adds redundant samples: the number of independent samples remains .
Power
Time integration relies on copying the ligand-binding state of the receptor into the chemical modification states of the readout molecules (Mehta and Schwab, 2012; Govern and Ten Wolde, 2014a). This copy process correlates the state of the receptor with that of the readout, which requires work input (Ouldridge et al., 2017).
The free-energy provided by the fuel turnover drives the readout around the cycle of modification and demodification (Figure 1). The rate at which the fuel molecules do work is the power , and the total work performed during the integration time is . This work is spent on taking samples of receptor molecules that are bound to ligand because only they can modify the readout. The total number of effective samples of ligand-bound receptors during is (Equation 7), which means that the work per effective sample of a ligand-bound receptor is (Govern and Ten Wolde, 2014a).
To understand how energy limits the sensing precision, we can distinguish between two limiting regimes (Govern and Ten Wolde, 2014a). When , the quality parameter (Equation 7) and the work per sample of a ligand-bound receptor is (Govern and Ten Wolde, 2014a). In this irreversible regime, the SNR bound is given by Equation 8 with . The power limits the sensing accuracy not because it limits the reliability of each sample but because it limits the rate at which the receptor is sampled.
When , the system enters the quasi-equilibrium regime in which the quality parameter (see Equation 7, noting that in the optimal system ). The sensing bound is now given by Equation 8 with , which is larger than in the irreversible regime (where ). The quasi-equilibrium regime minimizes the sensing error for a given power constraint (Figure 3b) because this regime maximizes the number of effective measurements per work input (Govern and Ten Wolde, 2014a).
While the sensing precision for a given power and time constraint is higher in the quasi-reversible regime, more readout molecules are required to store the concentration measurements in this regime. Noting that the flux , it follows that in the irreversible regime () the number of readout molecules consuming energy at a rate is
while in the quasi-equilibrium regime () it is
Since in the quasi-equilibrium regime , .
Equation 8 shows that the sensing precision is fundamentally bounded not by the work , as observed for static signals (Govern and Ten Wolde, 2014a), but rather by the power and the integration time separately such that an optimal integration time emerges. Figure 3f shows how depends on . Since the system cannot sense without any readout molecules, in the low-power regime the system maximizes subject to the power constraint (see Equations 10 and 11) by making as large as possible, which is the signal correlation time – increasing further would average out the signal itself. As is increased, rises and the sampling error decreases. When the sampling error becomes comparable to the dynamical error (Equation 6), the system starts to trade a further reduction in the sampling error for a reduction in the dynamical error by decreasing . The sampling error and dynamical error are now reduced simultaneously by increasing and decreasing . This continues until the sampling interval becomes comparable to the receptor correlation time , as marked by the yellow bar. Beyond this point, and the sampling error is no longer limited by but rather by since bounds the number of independent samples per receptor, . The system has entered the Berg–Purcell regime, where is determined by the trade-off between the dynamical error and the sampling error as set by the maximum number of independent samples, (Figure 3c).
Optimal design
In sensing time-varying signals, a trade-off between time averaging and signal tracking is inevitable. Moreover, the optimal integration time depends on which resource is limiting, being zero when is limiting and finite when or is limiting (Figure 3). It is therefore not obvious whether these sensing systems still obey the optimal resource allocation principle as observed for systems sensing static concentrations (Govern and Ten Wolde, 2014a).
However, Equation 8 shows that when for a given integration time , , the bounds on the sensing precision as set by, respectively, the number of receptors , the number of readout molecules , and the power are equal. Each of these resources is then equally limiting sensing and no resource is in excess. We thus recover the optimal resource allocation principle:
Irrespective of whether the concentration fluctuates in time, the number of independent concentration measurements at the receptor level is , which in an optimally designed system also equals the number of readout molecules and the energy that are both necessary and sufficient to store these measurements reliably.
The design principle (Equation 12) predicts that there exists a driving force that optimizes the trade-off between the number of samples and their accuracy. Noting that reveals that the principle (Equation 12) specifies for the optimal system in which and via the equation , where is defined in Equation 7. A numerical inspection shows that to a good approximation the solution of this equation is precisely given by the crossover from the quasi-equilibrium regime to the irreversible one: . This can be understood by noting that in the quasi-equilibrium regime can, for a given power and time constraint, be reduced by increasing (Equation 11) without compromising the sensing precision (Equation 8 with ); in this regime, increasing increases the reliability of each sample, and a smaller number of more reliable samples precisely compensates for a larger number of less reliable ones. Yet, when becomes larger than , the system enters the irreversible regime. Here, corresponding to a given and constraint still decreases with (Equation 10), but the sensing error now increases (Equation 8 with ) because each sample has become (essentially) perfect in this regime – hence, the samples’ accuracy cannot (sufficiently) increase further to compensate for the reduction in the sampling rate .
Equation 12 holds for any integration time , yet it does not specify . The cell membrane is highly crowded, and many systems employ time integration (Berg and Purcell, 1977; Bialek and Setayeshgar, 2005; Govern and Ten Wolde, 2014a). This suggests that these systems employ time integration and accept the signal distortion that comes with it simply because there is not enough space on the membrane to increase . Our theory then allows us to predict the optimal integration time based on the premise that is limiting. As Equation 8 reveals, in this limit does not only depend on but also on , , and . The optimal design of the system is then given by Equation 12 but with given by :
This design principle maximizes for a given number of receptors the sensing precision and minimizes the number of readout molecules and power needed to reach that precision.
Comparison with experiment
To test our theory, we turn to the chemotaxis system of E. coli. This system contains a receptor that forms a complex with the kinase CheA. This complex, which is coarse-grained into R (Govern and Ten Wolde, 2014a), can bind the ligand L and activate the intracellular messenger protein CheY (x) by phosphorylating it. Deactivation of CheY is catalyzed by CheZ, the effect of which is coarse-grained into the deactivation rate. This push–pull network allows E. coli to measure the current concentration, and the relaxation time of this network sets the integration time for the receptor (Sartori and Tu, 2011). The system also exhibits adaptation on longer timescales due to receptor methylation and demethylation. The push–pull network and the adaptation system together allow the cell to measure concentration gradients via a temporal derivative, taking the difference between the current concentration and the past concentration as set by the adaptation time (Segall et al., 1986). A lower bound for the error in the estimate of this difference is given by the error in the estimate of the current concentration, the central quantity of our theory. Here, we ask how accurately E. coli can estimate the latter and whether the sensing precision is sufficient to determine whether during a run the concentration has changed.
Our theory predicts that if the number of receptors is limiting then the optimal integration time is given by minimizing Equation 8 with . The number of receptor–CheA complexes depends on the growth rate and varies between and (Li and Hazelbauer, 2004). The receptor correlation time for the binding of aspartate to the Tar receptor can be estimated from the measured dissociation constant (Vaknin and Berg, 2007) and the association rate (Danielson et al., 1994), (Govern and Ten Wolde, 2014a). The timescale of the input fluctuations is set by the typical run time, which is on the order of a few seconds, (Berg and Brown, 1972; Taute et al., 2015).
This leaves one parameter to be determined, . This is set by the spatial ligand–concentration profile and the typical length of a run. We have a good estimate of the latter. In shallow gradients, it is on the order of (Berg and Brown, 1972; Taute et al., 2015; Jiang et al., 2010; Flores et al., 2012); specifically, Figure 4 of Taute et al., 2015 shows that the typical run times are 1–2 s while the typical run speeds are , yielding a run length on the order of indeed 50 µm. We do not know the spatial concentration profiles that E. coli has experienced during its evolution. We can however get a sense of the scale by considering an exponential ligand–concentration gradient. For a profile with length scale x0, the relative change in the signal over the length of a run is . We consider the range , where corresponds to shallow gradients with in which cells move with a constant drift velocity (Shimizu et al., 2010; Flores et al., 2012).
Figure 4a shows that as the gradient becomes steeper and increases the optimal integration time decreases. This can be understood by noting that the relative importance of the dynamical error compared to the sampling error scales with (Equation 6). Shallow ingredients thus allow for a larger integration time while steep gradients necessitate a shorter one.
Experiments indicate that the relaxation rate of CheY is for the attractant response and for the repellent response (Sourjik and Berg, 2002), such that the integration time (Sourjik and Berg, 2002; Govern and Ten Wolde, 2014a). Figure 4a shows that this integration time is optimal for detecting shallow gradients. Our theory thus predicts that the E. coli chemotaxis system has been optimized for sensing shallow gradients.
To navigate, the cells must be able to resolve the signal change over a run. During a run of duration , the system performs independent concentration measurements. The effective error for these measurements is the instantaneous sensing error divided by the number of independent measurements . Hence, the SNR for these concentration measurements is .
Figure 4b shows that our theory predicts that when , the shallowest gradient that cells can resolve, defined by , is , corresponding to , while when , and . The shallowest gradient is thus on the order of . Shimizu et al., 2010 show that E. coli cells are indeed able to sense such very shallow gradients: Figure 2A of Shimizu et al., 2010 shows that E. coli cells can detect exponential up ramps with rate ; using , where is the run speed (Jiang et al., 2010), this corresponds to . Importantly, the predictions of our theory (Figure 4) concern the shallowest gradient that the system with the optimal integration time can resolve. These observations indicate that the optimal integration time is not only sufficient to make navigation in these very shallow gradients possible but also necessary.
Figure 4 also shows that decreases as the number of receptor–CheA complex, , increases because the latter allows for more instantaneous measurements, reducing the need for time integration (Figure 3c). Interestingly, the data of Li and Hazelbauer, 2004 shows that the copy numbers of the chemotaxis proteins vary with the growth rate. Clearly, it would be of interest to directly measure the response time in different strains under different growth conditions.
Discussion
Here, we have integrated ideas from Tostevin and ten Wolde, 2010; Hilfinger and Paulsson, 2011; and Bowsher et al., 2013 on information transmission via time-varying signals with the sampling framework of Govern and Ten Wolde, 2014a to develop a unified theory of cellular sensing. The theory is founded on the concept of the dynamic input–output relation . It allows us to develop the idea that the cell employs the readout system to estimate the average receptor occupancy over the past integration time and then exploits the mapping to estimate the current ligand concentration L from . The theory reveals that the error in the estimate of L depends on how accurately the cell samples the receptor state to estimate , and on how much , which is determined by the concentration in the past , reflects the current ligand concentration. These two distinct sources of error give rise to the sampling error and dynamical error in Equation 6, respectively.
While the system contains no less than 11 parameters, Equation 6 provides an intuitive expression for the sensing error in terms of collective variables that have a clear interpretation. The dynamical error depends only on the timescales in the problem, most notably . The sampling error depends on how accurately the readout system estimates , which is determined by the number of receptor samples, their independence, and their accuracy; yet it also depends on via the dynamic gain, which determines how the error in the estimate of propagates to that of L. The trade-off between the sampling error and dynamical error yields an optimal integration time.
Our study reveals that the optimal integration time depends in a non-trivial manner on the design of the system. When the number of readout molecules is smaller than the number of receptors , time integration is not possible and the optimal system is an instantaneous responder with . When the power , rather than , is limiting, is determined by the trade-off between the sampling error and dynamical error. In both scenarios, however, one resource, or , is limiting the sensing precision. In an optimally designed system, all resources are equally limiting so that no resource is wasted. This yields the resource allocation principle (Equation 12), first identified in Govern and Ten Wolde, 2014a, for sensing static concentrations. The reason it can be generalized to time-varying signals is that the principle concerns the optimal design of the readout system for estimating the receptor occupancy over a given integration time , which holds for any type of input: the number of independent concentration measurements at the receptor level is , irrespective of how the input varies, and in an optimally designed system this also equals the number of readout molecules and energy to store these measurements reliably. We thus expect that the design principle also holds for systems that sense signals that vary more strongly in time (Mora and Nemenman, 2019).
While the allocation principle Equation 12 holds for any , it does not specify the optimal integration time . However, our theory predicts that if the number of receptors is limiting, then there exists a that maximizes the sensing precision for that (Equation 8 with ). Via the allocation principle Equation 13, and then together determine the minimal number of readout molecules and power to reach that precision. The resource allocation principle, together with the optimal integration time, thus completely specifies the optimal design of the sensing system.
Applying our theory to the E. coli chemotaxis system shows that this system not only obeys the resource allocation principle (Govern and Ten Wolde, 2014a) but also that the predicted optimal integration time to measure shallow gradients is in agreement with that measured experimentally (Figure 4a). This is remarkable because there is not a single fit parameter in our theory. Moreover, Figure 4b shows that the optimal integration time is not only sufficient to enable the sensing of these shallow gradients but also necessary. This is interesting because the sensing precision could also be increased by increasing the number of receptors, readout molecules, and energy devoted to sensing – but this would be costly. Our results thus demonstrate not only that the chemotaxis system obeys the design principles as revealed by our theory but also that there is a strong selection pressure to design sensing systems optimally, that is, to maximize the sensing precision given the resource constraints.
Our theory is based on a Gaussian model and describes the optimal sensing system that minimizes the mean square error in the estimate of the ligand concentration (see Equation 1). The latter is precisely the performance criterion of Wiener–Kolmogorov (Extrapolation, 1950; Kolmogorov, 1992) and Kalman, 1960 filtering theory, which, moreover, become exact for systems that obey Gaussian statistics. In fact, since our system (including the input signal) is stationary, they predict the same optimal filter, which is an exponential filter for signals that are memoryless. The signals studied here belong to this class, and the push–pull network forms an exponential filter (Hinczewski and Thirumalai, 2014; Becker et al., 2015). This underscores the idea that our theory gives a complete description, in terms of all the required resources, for the optimal design of cellular sensing systems that need to estimate this type of signals. Furthermore, because our model is Gaussian, the goal of minimizing the mean-square error in the estimate of the input signal is equivalent to maximizing the mutual information between the input (the ligand concentration) and the output (the readout ) (Becker et al., 2015).
In recent years, filtering theories and information theory have been applied increasingly to neuronal and cellular systems (Laughlin, 1981; Brenner et al., 2000; Fairhall et al., 2001; Andrews et al., 2006; Ziv et al., 2007; Nemenman et al., 2008; Cheong et al., 2011; Nemenman, 2012; Hinczewski and Thirumalai, 2014; Becker et al., 2015; Husain et al., 2019; Tkacik et al., 2008; Tkačik and Walczak, 2011; Dubuis et al., 2013; Monti and Wolde, 2016; Monti et al., 2018a). A key concept in these theories is that optimal sensing systems match the response to the statistics of the input. When the noise is weak, maximizing the entropy of the output distribution becomes paramount, which entails matching the shape of the input–output relation to the shape of the input distribution to generate a flat output distribution (Laughlin, 1981; Tkacik et al., 2008; Monti et al., 2018a). Yet, when the noise is large, the optimal response is also shaped by the requirement to tame the propagation of noise in the input signal (Andrews et al., 2006; Hinczewski and Thirumalai, 2014; Becker et al., 2015; Monti et al., 2018a; Monti et al., 2018b; Mora and Nemenman, 2019) or to lift the signal above the intrinsic noise in the response system (Tostevin and ten Wolde, 2010; Bowsher et al., 2013). In Appendix 3, we show that estimating the concentration from is equivalent to that via readout . This makes it possible to connect our sampling framework, which is based on , to filtering and information theory, which are based on . In particular, we show in this appendix how the optimal integration and dynamic gain can be understood from these ideas on matching the response to the input. We also briefly discuss in Appendix 3 the concepts from information theory that are beyond the scope of the Gaussian model considered here.
Yet, our discrete sampling framework gives a detailed description of how the optimal design of sensing systems depends on the statistics of the input signal in terms of all the required cellular resources: protein copies, time, and energy. In an optimal system, each receptor is sampled once every receptor–ligand correlation time , , and the number of samples per receptor is . The optimal integration time for a given is determined by the trade-off between the age of the samples and the number required for averaging the receptor state. When the input varies more rapidly, the samples need to be refreshed more regularly: to keep the dynamical error and the dynamic gain constant, must decrease linearly with (see Equation 6). Yet, only decreasing would inevitably increase the sampling error in estimating the receptor occupancy because the sampling interval would become smaller than , creating redundant samples. To keep the sensing precision constant, the number of receptors needs to be raised with , such that the sampling interval remains of order and the decrease in the number of samples per receptor, , is precisely compensated for by the increase in . The total number of independent concentration measurements, , and hence the number of readout molecules to store these, does indeed not change. In contrast, the required power rises (Equation 12): each receptor molecule is sampled each at , and the increase in raises the sampling rate . Our theory thus predicts that when the input varies more rapidly the number of receptors and the power must rise to maintain a required sensing precision, while the number of readout molecules does not.
The fitness benefit of a sensing system does not only depend on the sensing precision but also on the energetic cost of maintaining and running the system. In principle, the cell can reduce the sensing error arbitrarily by increasing and decreasing . Our resource allocation principle (Equation 12) shows that then not only the number of readout molecules needs to be raised but also the power. Clearly, improving the sensing precision comes at a cost: more copies of the components of the sensing system need to be synthesized every cell cycle, and more energy is needed to run the system. Our theory (i.e., Equation 6) makes it possible to derive the Pareto front that quantifies the trade-off between the maximal sensing precision and the cost of making the sensing system (see Figure 5). Importantly, the design of the optimal system at the Pareto front obeys, to a good approximation, our resource allocation principle (Equation 12). This is because this principle specifies the optimal ratios of , , , and given the input statistics, and these ratios are fairly insensitive to the costs of the respective resources: resources that are in excess cannot improve sensing and are thus wasted, no matter how cheap they are. It probably explains why our theory, without any fit parameters, not only predicts the integration time that allows E. coli to sense shallow gradients (Figure 4) but also the number of receptor and readout molecules (Govern and Ten Wolde, 2014a).
In our study, we have limited ourselves to a canonical push–pull motif. Yet, the work of Govern and Ten Wolde, 2014a indicates that our results hold more generally, pertaining also to systems that employ cooperativity, negative or positive feedback, or multiple layers, as the MAPK cascade. While multiple layers and feedback change the response time, they do not make time integration more efficient in terms of readout molecules or energy (Govern and Ten Wolde, 2014a). And provided it does not increase the input correlation time (Skoge et al., 2011; Ten Wolde et al., 2016), cooperative ligand binding can reduce the sensing error per sample, but the resource requirements in terms of readout molecules and energy per sample do not change (Govern and Ten Wolde, 2014a). In all these systems, time integration requires that the history of the receptor is stored, which demands protein copies and energy.
Lastly, in this article we have studied the resource requirements for estimating the current concentration via the mechanism of time integration. However, to understand how E. coli navigates in a concentration gradient, we do not only have to understand how the system filters the high-frequency ligand-binding noise via time averaging but also how on longer timescales the system adapts to changes in the ligand concentration (Sartori and Tu, 2011). This adaptation system also exhibits a trade-off between accuracy, speed, and power (Lan et al., 2012; Sartori and Tu, 2015). Intriguingly, simulations indicate that the combination of sensing and adaptation allows E. coli not only to accurately estimate the current concentration but also the future ligand concentration (Becker et al., 2015). It will be interesting to see whether an optimal resource allocation principle can be formulated for systems that need to predict future ligand concentrations.
Materials and methods
Methods are described in Appendices 1–3. Appendix 1 derives the central result of our article (Equation 6). Appendix 2 derives the fundamental resources and the corresponding sensing limits (Equations 8 and 9). Appendix 3 describes how the optimal gain and integration time can be understood using ideas from filtering and information theory.
Appendix 1
Signal-to-noise ratio
Here, we provide the derivation of the central result of this article, Equation 6 of the main text. The derivation starts from the SNR, given in Equation 2. Here, is the width of the input distribution, while is the error in the estimate of the concentration. The latter is derived from the dynamic input–output relation , which is the mapping between the average receptor occupancy over the past integration time and the current ligand concentration L (see Figure 2). Concretely, the error is given by Equation 1, where is the error in the estimate of the average receptor occupancy over the past integration time and is the dynamic gain, which is the slope of the dynamic input–output relation . Below, we first derive the dynamic gain and then the error in the estimate of the receptor occupancy .
Dynamic input–output relation
The dynamic input–output relation is the average receptor occupancy over the past integration time , given that the current ligand concentration . The cell estimates via its receptor readout system, which is a device that takes samples of the receptor: the readout molecules at time t constitute samples of the ligand-binding state of the receptor at earlier sampling times ti (see Figure 2). More specifically, the cell estimates from the number of active readout molecules :
where is the average of the number of samples N taken during the integration time . Hence, the dynamic input–output relation is
where is the receptor occupancy at time ti, E denotes the expectation over the sampling times ti, and denotes an average over receptor–ligand binding noise and the subensemble of ligand trajectories that each end at (see Figure 2c); the quantity is indeed the average receptor occupancy at time ti, given that the ligand concentration at time t is . Importantly, the receptor samples can also decay via the deactivation of . Taking this into account, the probability that a readout molecule at time t provides a sample of the receptor at an earlier time ti is (Govern and Ten Wolde, 2014a). Averaging the receptor occupancy over the sampling times ti then yields
Dynamic gain
When the current ligand concentration deviates from its mean by , then deviates on average from its mean p (the average receptor occupancy over all ) by
Here, E denotes again the expectation over the sampling times ti, and is the average deviation in the receptor occupancy at time ti from its mean p, given that the ligand concentration at time t is (see Figure 2c). We can compute it within the linear-noise approximation (Gardiner, 2009):
where and is the average ligand concentration at time , given that the ligand concentration at time t is . The latter is given by Bowsher et al., 2013
Combining Equations 17–19 yields the following expression for the average change in the average receptor occupancy , given that the ligand at time t is :
Hence, the dynamic gain is
The dynamic gain is the slope of the dynamic input–output relation (see Figure 2a). It yields the average change in the receptor occupancy over the past integration time when the change in the ligand concentration at time t is . It depends on all the timescales in the problem and only reduces to the static gain when the integration time and the receptor correlation time are both much shorter than the ligand correlation time . The dynamic gain determines how much an error in the estimate of propagates to the estimate of .
Error in receptor occupancy
We can derive the variance in the estimate of the receptor occupancy over the past integration time , , directly from Equation 14 for the system in the irreversible limit (Malaguti and Ten Wolde, 2019). While this derivation is illuminating, it is also lengthy. For the fully reversible system studied here, we follow a simpler route. Since the average number of samples over the integration time is constant, it follows from Equation 14 that
where is the variance in the number of phosphorylated readout molecules, conditioned on the signal at time t being . The conditional variance (Tostevin and ten Wolde, 2010)
is the full variance of minus the variance that is due to the signal variations, given by the dynamic gain from L to times the signal variance .
The full variance of the readout in Equation 25 can be obtained from the linear-noise approximation (Gardiner, 2009), see Malaguti and Ten Wolde, 2019:
In this expression, is the inverse of the receptor correlation time is the probability that a receptor is bound to ligand; ; is the inverse of the integration time ; is the fraction of phosphorylated readout; and is the total flux around the cycle of readout activation and deactivation divided by the total number of ligand-bound receptors: it is the rate at which each receptor is sampled, be it ligand bound or not. For what follows below, we note that the quality parameter .
To get from Equations 24 and 25, we need not only (Equation 26) but also the average number of samples and the dynamic gain . The average number of samples taken during the integration time is , and the effective number of reliable samples is . Since , where is the average number of active readout molecules for a given input and is a constant independent of L, it follows that
with the dynamic gain from L to , given by Equation 22. Equation 27 can be verified via another route that does not rely on the sampling framework because we also know that (Tostevin and ten Wolde, 2010), where the co-variance can be obtained from the linear-noise approximation (Malaguti and Ten Wolde, 2019; Gardiner, 2009). Combining Equations 24–27 yields
This can be rewritten using the expression for the fraction of independent samples, which, assuming that , is , with the effective spacing between the samples (Govern and Ten Wolde, 2014a):
Here, is the sampling error in the estimate of (Malaguti and Ten Wolde, 2019); it is a statistical error, which arises from the finite cellular resources to sample the state of the receptor, protein copies, time, and energy (see Figure 2b). The other contribution, , is the dynamical error in the estimate of (Malaguti and Ten Wolde, 2019); it is a systematic error that arises from the input dynamics and only depends on the average receptor occupancy and the timescales of the input, receptor, and readout (see Figure 2c); it neither depends on the number of protein copies nor on the energy necessary to sample the receptor.
Final result: SNR
Combining Equations 29 and 22 with Equation 3 yields the principal result of our work (Equation 6) of the main text.
Appendix 2
Fundamental resources
To identify the fundamental resources limiting the sensing accuracy and derive the corresponding sensing limits (Equations 8 and 9), it is helpful to rewrite the SNR in terms of collective variables that illuminate the cellular resources. For that, we start from Equation 6 of the main text and split the first term on the right-hand side and exploit the expression for the effective number of independent samples with . We then sum up the last two terms on the right-hand side and use that :
The second term in between the square brackets describes the contribution to the sensing error that comes from the stochasticity in the concentration measurements at the receptor level. The first term in between the square brackets, the coding noise, describes the contribution that arises in storing these measurements into the readout molecules.
From Equation 30, the fundamental resources and the corresponding sensing limits (Equations 8 and 9) can be derived. Specifically, when the number of receptors and their integration are limiting, the coding noise in Equation 30 is zero; exploiting that typically and that the contribution to the sensing error from the receptor input noise is minimized for , this yields Equation 8 with . When the number of readout molecules is limiting, the receptor input noise is zero and ; noting that and that the contribution from the coding noise is minimized when and , and again exploiting that , this yields Equation 8 with . When the power is limiting, then the receptor input noise is (again) zero. The coding noise is minimized for a given power constraint when , but two regimes can be distinguished based on the total free-energy drop . When , the system is in the irreversible regime and (see Equation 7); Equation 30 shows that the error is then bounded by Equation 8 with , using and . Yet, the sensing error is minimized in the quasi-equilibrium regime, where and , yielding Equation 8 with .
Appendix 3
The optimal gain and optimal integration time
The theory of the main text (Equation 6) is based on the idea that the cell uses its push–pull network to estimate the receptor occupancy from which the current ligand concentration L is then inferred by inverting the dynamic input–output relation . Yet, as we show here, this framework is equivalent to the idea that the cell estimates the concentration from the output , using the dynamic input–output relation . Here, we use this observation to analyze our system using ideas from filtering and information theory. But first we demonstrate this correspondence.
To show that estimating the concentration from is equivalent to that from estimating it from , we first note that because the average number of samples is constant, while the gain from L to is . Consequently, the absolute error in estimating the concentration via , , is the same as that of Equation 1: because the instantaneous number of active readout molecules reflects the average receptor occupancy over the past , estimating the ligand concentration from is no different from inferring it from the average receptor occupancy .
To make the connection with information and filtering theory, we note that in our Gaussian model the conditional distribution of given is given by Tostevin and ten Wolde, 2010
where is the average value of given that , and is the variance of this distribution (see also Equation 25).
The relative error, the inverse of the SNR (see Equation 2), is
As mentioned in the main text, the SNR also yields the mutual information between the input L and output (Tostevin and ten Wolde, 2010).
The notion of an optimal integration time or optimal dynamic gain is well known from filtering and information theory (Andrews et al., 2006; Hinczewski and Thirumalai, 2014; Becker et al., 2015; Monti et al., 2018a; Monti et al., 2018b; Mora and Nemenman, 2019). To elucidate the optimal gain and integration time in our system, we combine the above equation with Equations 25 and 26 to write the relative error as
where is the static gain from to . Written in this form, the trade-offs in maximizing the mutual information (and minimizing the relative error in estimating the concentration) become apparent: increasing the dynamic gain by decreasing the integration time raises the slope of the input–output relation , which helps to lift the transmitted signal above the intrinsic binomial switching noise of the readout, . Also, the dynamical error is minimized by minimizing and maximizing . Yet, for the second term, which describes how noise in the input signal arising from receptor switching, , is propagated to the output , there exists an optimal integration time that minimizes this term: while decreasing increases the dynamic gain, which helps to raise the signal above the noise, it also impedes time averaging of this switching noise, described by the factor .
The mutual information is , with the entropy of the marginal output distribution and the entropy of the output distribution conditioned on the input. Hence, information theory shows that in the weak noise limit, information transmission is optimal when the entropy of the output distribution is maximized (Laughlin, 1981; Tkacik et al., 2008). Our system obeys this principle. Since the dynamic gain , the amplification of the signal rises with and . Since the standard deviation of the noise added to the transmitted signal coming from the stochastic receptor and readout activation scales with and , respectively, it is clear that the SNR increases with and . In the limit that , the relative error is only set by the dynamical error, which can be reduced to zero by , exploiting that typically . This is the weak-noise limit in which the mutual information is maximized by maximizing the entropy of the output distribution . Indeed, corresponds to maximizing the gain, which maximizes the width of the output distribution, in this limit equal to (see Equation 25), and thereby the entropy of the output distribution .
Finally, we note that our Gaussian model is linear such that the central control parameter, besides protein copies and energy, is the integration time or the dynamic gain, which sets the slope of the linear input–output relation. While Wiener–Kolmogorov and Kalman filtering are exact only for these Gaussian models, information theory also applies to non-linear systems with non-Gaussian statistics. It has been used to show that neuronal systems (Laughlin, 1981; Brenner et al., 2000; Fairhall et al., 2001; Nemenman et al., 2008; Tkacik et al., 2010), signaling and gene networks (Segall et al., 1986; Tkacik et al., 2008; Tkačik and Walczak, 2011; Nemenman, 2012; Dubuis et al., 2013), and circadian systems (Monti and Wolde, 2016; Monti et al., 2018a) can maximize information transmission by optimizing the shape of the input–output relation (Laughlin, 1981; Brenner et al., 2000; Fairhall et al., 2001; Tkacik et al., 2008; Monti et al., 2018a); by desensitization, that is, adapting the output to the mean input via incoherent feedforward or negative feedback (Segall et al., 1986); by gain control, that is, adapting the output to the variance of the input by capitalizing on a steep response function and temporal correlations in the input (Nemenman, 2012); by removing coding redundancy via temporal decorrelation (Nemenman et al., 2008); by optimizing the tiling of the output space via the topology of the network (Tkačik and Walczak, 2011; Dubuis et al., 2013); or by exploiting cross-correlations between the signals (Tkacik et al., 2010; Monti and Wolde, 2016).
Data availability
All data generated or analysed during this study are included in the manuscript and supporting files.
References
-
BookIntroduction to Systems Biology: Design Principles of Biological NetworksBoca Raton, FL: CRC press.https://doi.org/10.1016/j.mbs.2008.07.002
-
Optimal noise filtering in the chemotactic response of Escherichia coliPLOS Computational Biology 2:e154.https://doi.org/10.1371/journal.pcbi.0020154
-
Optimal prediction by cellular signaling networksPhysical Review Letters 115:258103.https://doi.org/10.1103/PhysRevLett.115.258103
-
Physics of chemoreceptionBiophysical Journal 20:193–219.https://doi.org/10.1016/S0006-3495(77)85544-6
-
Physical limits to biochemical signalingPNAS 102:10040–10045.https://doi.org/10.1073/pnas.0504321102
-
The fidelity of dynamic signaling by noisy biomolecular networksPLOS Computational Biology 9:e1002965.https://doi.org/10.1371/journal.pcbi.1002965
-
Bicoid gradient formation mechanism and dynamics revealed by protein lifetime analysisMolecular Systems Biology 14:e8355.https://doi.org/10.15252/msb.20188355
-
Maximum likelihood and the single receptorPhysical Review Letters 103:158101.https://doi.org/10.1103/PhysRevLett.103.158101
-
BookExtrapolation, Interpolation, and Smoothing of Stationary Time Series: With Engineering ApplicationsMIT Press.
-
Fundamental limits to collective concentration sensing in cell populationsPhysical Review Letters 118:078101.https://doi.org/10.1103/PhysRevLett.118.078101
-
Signaling noise enhances chemotactic drift of E. coliPhysical Review Letters 109:148101.https://doi.org/10.1103/PhysRevLett.109.148101
-
BookStochastic Methods: A Handbook for the Natural and Social SciencesBerlin: Springer-Verlag.
-
Fundamental limits on sensing chemical concentrations with linear biochemical networksPhysical Review Letters 109:218103.https://doi.org/10.1103/PhysRevLett.109.218103
-
Energy dissipation and noise correlations in biochemical sensingPhysical Review Letters 113:258102.https://doi.org/10.1103/PhysRevLett.113.258102
-
Physical limits on cellular sensing of spatial gradientsPhysical Review Letters 105:048104.https://doi.org/10.1103/PhysRevLett.105.048104
-
Quantitative modeling of Escherichia coli chemotactic motion in environments varying in space and timePLOS Computational Biology 6:e1000735.https://doi.org/10.1371/journal.pcbi.1000735
-
The Berg-Purcell limit revisitedBiophysical Journal 106:976–985.https://doi.org/10.1016/j.bpj.2013.12.030
-
A new approach to linear filtering and prediction problemsJournal of Basic Engineering 82:35–45.https://doi.org/10.1115/1.3662552
-
BookProbability theory and mathematical statisticsIn: Watanabe S, Prokhorov J. V, editors. Selected Works of A. N. Kolmogorov. Netherlands: Springer Science & Business Media. pp. 8–14.https://doi.org/10.1007/BFb0078455
-
The energy-speed-accuracy tradeoff in sensory adaptationNature Physics 8:422–428.https://doi.org/10.1038/nphys2276
-
Thermodynamics of statistical inference by cellsPhysical Review Letters 113:148103.https://doi.org/10.1103/PhysRevLett.113.148103
-
A simple coding procedure enhances a neuron's information capacityZeitschrift Für Naturforschung C 36:910–912.https://doi.org/10.1515/znc-1981-9-1040
-
Cellular stoichiometry of the components of the chemotaxis signaling complexJournal of Bacteriology 186:3687–3694.https://doi.org/10.1128/JB.186.12.3687-3694.2004
-
Feedback between motion and sensation provides nonlinear boost in run-and-tumble navigationPLOS Computational Biology 13:e1005429.https://doi.org/10.1371/journal.pcbi.1005429
-
Energetic costs of cellular computationPNAS 109:17978–17982.https://doi.org/10.1073/pnas.1207814109
-
Optimal entrainment of circadian clocks in the presence of noisePhysical Review E 97:032405.https://doi.org/10.1103/PhysRevE.97.032405
-
Robustness of clocks to input noisePhysical Review Letters 121:078101.https://doi.org/10.1103/PhysRevLett.121.078101
-
The accuracy of telling time via oscillatory signalsPhysical Biology 13:035005–035014.https://doi.org/10.1088/1478-3975/13/3/035005
-
Physical limit to concentration sensing in a changing environmentPhysical Review Letters 123:198101.https://doi.org/10.1103/PhysRevLett.123.198101
-
Limits of sensing temporal concentration changes by single cellsPhysical Review Letters 104:248101.https://doi.org/10.1103/PhysRevLett.104.248101
-
Neural coding of natural stimuli: information at sub-millisecond resolutionPLOS Computational Biology 4:e1000025.https://doi.org/10.1371/journal.pcbi.1000025
-
Gain control in molecular information processing: lessons from neurosciencePhysical Biology 9:026003–026008.https://doi.org/10.1088/1478-3975/9/2/026003
-
Thermodynamics of computational copying in biochemical systemsPhysical Review X 7:021004.https://doi.org/10.1103/PhysRevX.7.021004
-
Receptor noise and directional sensing in eukaryotic chemotaxisPhysical Review Letters 100:228101.https://doi.org/10.1103/PhysRevLett.100.228101
-
Noise filtering strategies in adaptive biochemical signaling networks: application to E. coli chemotaxisJournal of Statistical Physics 142:1206–1217.https://doi.org/10.1007/s10955-011-0169-z
-
Free energy cost of reducing noise while maintaining a high sensitivityPhysical Review Letters 115:118102.https://doi.org/10.1103/PhysRevLett.115.118102
-
Dynamics of cooperativity in chemical sensing among cell-surface receptorsPhysical Review Letters 107:178101.https://doi.org/10.1103/PhysRevLett.107.178101
-
Fundamental limits to cellular sensingJournal of Statistical Physics 162:1395–1424.https://doi.org/10.1007/s10955-015-1440-5
-
Information transmission in genetic regulatory networks: a reviewJournal of Physics: Condensed Matter 23:153102.https://doi.org/10.1088/0953-8984/23/15/153102
-
Mutual information between input and output trajectories of biochemical networksPhysical Review Letters 102:218101.https://doi.org/10.1103/PhysRevLett.102.218101
-
Mutual information in time-varying biochemical systemsPhysical Review E 81:061917.https://doi.org/10.1103/PhysRevE.81.061917
-
Physical responses of bacterial chemoreceptorsJournal of Molecular Biology 366:1416–1423.https://doi.org/10.1016/j.jmb.2006.12.024
-
Quantifying noise levels of intercellular signalsPhysical Review E 75:061905.https://doi.org/10.1103/PhysRevE.75.061905
Article and author information
Author details
Funding
Nederlandse Organisatie voor Wetenschappelijk Onderzoek
- Giulia Malaguti
- Pieter Rein ten Wolde
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
Acknowledgements
We wish to acknowledge Bela Mulder, Tom Shimizu, and Tom Ouldridge for many fruitful discussions and a careful reading of the manuscript. This work is part of the research program of the Netherlands Organisation for Scientific Research (NWO) and was performed at the research institute AMOLF.
Copyright
© 2021, Malaguti and ten Wolde
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
-
- 1,474
- views
-
- 330
- downloads
-
- 13
- citations
Views, downloads and citations are aggregated across all versions of this paper published by eLife.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading
-
- Physics of Living Systems
Many proteins have been recently shown to undergo a process of phase separation that leads to the formation of biomolecular condensates. Intriguingly, it has been observed that some of these proteins form dense droplets of sizeable dimensions already below the critical concentration, which is the concentration at which phase separation occurs. To understand this phenomenon, which is not readily compatible with classical nucleation theory, we investigated the properties of the droplet size distributions as a function of protein concentration. We found that these distributions can be described by a scale-invariant log-normal function with an average that increases progressively as the concentration approaches the critical concentration from below. The results of this scaling analysis suggest the existence of a universal behaviour independent of the sequences and structures of the proteins undergoing phase separation. While we refrain from proposing a theoretical model here, we suggest that any model of protein phase separation should predict the scaling exponents that we reported here from the fitting of experimental measurements of droplet size distributions. Furthermore, based on these observations, we show that it is possible to use the scale invariance to estimate the critical concentration for protein phase separation.
-
- Computational and Systems Biology
- Physics of Living Systems
Explaining biodiversity is a fundamental issue in ecology. A long-standing puzzle lies in the paradox of the plankton: many species of plankton feeding on a limited variety of resources coexist, apparently flouting the competitive exclusion principle (CEP), which holds that the number of predator (consumer) species cannot exceed that of the resources at a steady state. Here, we present a mechanistic model and demonstrate that intraspecific interference among the consumers enables a plethora of consumer species to coexist at constant population densities with only one or a handful of resource species. This facilitated biodiversity is resistant to stochasticity, either with the stochastic simulation algorithm or individual-based modeling. Our model naturally explains the classical experiments that invalidate the CEP, quantitatively illustrates the universal S-shaped pattern of the rank-abundance curves across a wide range of ecological communities, and can be broadly used to resolve the mystery of biodiversity in many natural ecosystems.