Genetic control of encoding strategy in a foodsensing neural circuit
Abstract
Neuroendocrine circuits encode environmental information via changes in gene expression and other biochemical activities to regulate physiological responses. Previously, we showed that daf7 TGF$\beta $ and tph1 tryptophan hydroxylase expression in specific neurons encode food abundance to modulate lifespan in Caenorhabditis elegans, and uncovered cross and selfregulation among these genes (Entchev et al., 2015). Here, we now extend these findings by showing that these interactions between daf7 and tph1 regulate redundancy and synergy among neurons in food encoding through coordinated control of circuitlevel signal and noise properties. Our analysis further shows that daf7 and tph1 contribute to most of the foodresponsiveness in the modulation of lifespan. We applied a computational model to capture the general coding features of this system. This model agrees with our previous genetic analysis and highlights the consequences of redundancy and synergy during information transmission, suggesting a rationale for the regulation of these information processing features.
https://doi.org/10.7554/eLife.24040.001Introduction
Signaling pathways convey information about the environment, enabling organisms to generate appropriate physiological response to changing conditions (Gendron et al., 2015). We recently established that tph1 tryptophan hydroxylase expressed in ADF and NSM neurons and daf7 TGF$\beta $ expressed in ASI neurons in Caenorhabditis elegans transmit environmental information to physiology by modulating the response of lifespan to food (Entchev et al., 2015). Our previous analytical framework estimated the accuracy of tph1 and daf7 expression in decoding food input; however, it could not reveal the type of encoding strategy used by tph1 and daf7 within these neurons, nor could it quantify the contribution of these genes to lifespan modulation. Here, we applied information theory (Shannon, 1948) to address these issues. Information theory has been proposed as a general framework to characterize how biological signals are encoded and transmitted (Bowsher and Swain, 2014; Levchenko and Nemenman, 2014) and has been used to study information processing in the nervous system (Borst and Theunissen, 1999) as well as biochemical and genetic pathways (Cheong et al., 2011; Tkačik et al., 2015).
Groups of neurons can encode information redundantly or synergistically (Brenner et al., 2000; Puchalla et al., 2005). This form of informational redundancy is conceptually distinct from genetic redundancy. Redundant encoding systems replicate the same information in more than one neuron, analogous to a computer backup, which provides robustness to perturbations in single neurons at the expense of coding efficiency. In contrast, synergistic circuits encode more information than the sum of their component neurons, but this efficiency is vulnerable to disruptions in the constituent neurons. Redundancy and synergy have been defined using informationtheoretic measures (Averbeck et al., 2006; Schneidman et al., 2003), and both of these strategies for encoding information have been characterized in many neural and genetic circuits (Averbeck et al., 2006; Puchalla et al., 2005; Schneidman et al., 2011; Tkačik et al., 2015; Tkačik and Walczak, 2011).
Previously, we identified regulatory interactions among tph1 and daf7 that influence their coding accuracy (Entchev et al., 2015). Here, we show that crosstalk between daf7 and tph1 further affects the adoption of redundancy or synergy during discrimination between food levels. We found that the regulation of signaltonoise in gene expression underlies shifts between redundancy and synergy across genotypes. Finally, we use a computational model to explore the consequences of redundant and synergistic coding at the level of downstream targets.
Results and discussion
Information theory allows us to quantify the information encoded by daf7 and tph1 based on the overlap of their expression distributions (Figure 1A). By associating environmental stimuli (food level) and neuronal responses (gene expression) with the input and the output of a communication system, the encoding capacity of ASI, ADF, and NSM is given by the mutual information (MI) between gene expression responses (G) and food stimuli (F),
where $P\left(F\right)$ denotes the chances of encountering the food condition $F$, $P\left(G\rightF)$ is the response under each specific food level, and $P\left(G\right)$ is the average response across all the food stimuli (see Appendix and Figure 1—figure supplement 5). The MI measures the ability of the gene expression response to discriminate between food conditions.

Figure 1—source data 1
 https://doi.org/10.7554/eLife.24040.003

Figure 1—source data 2
 https://doi.org/10.7554/eLife.24040.004

Figure 1—source data 3
 https://doi.org/10.7554/eLife.24040.005

Figure 1—source data 4
 https://doi.org/10.7554/eLife.24040.006

Figure 1—source data 5
 https://doi.org/10.7554/eLife.24040.007
To define the redundancy of the system (Schneidman et al., 2003), we considered the difference between the sum of the information independently encoded by gene expression in the ADF, ASI, and NSM neurons, and the MI obtained from their combinatorial expression (Figure 1B):
Conceptually, redundancy occurs when the whole is less than the sum of parts ($R>0$), whereas synergy occurs when the whole is greater than the sum of parts ($R<0$) (Figure 1B).
This analysis revealed that ASI, ADF, and NSM neurons encode $\sim 0.9$ bits of information about food abundance in wildtype animals (Figure 1C), which is in the same range of information encoded by other biochemical pathways (Cheong et al., 2011), and it is consistent with the requirement for sensing the two states (boom or bust) experienced by C. elegans in the wild (Félix and Braendle, 2010). Approximately 40% of this information is encoded redundantly in wildtype animals (Figure 1D–E), consistent with the genetic evidence that tph1 and daf7 act in parallel pathways to modulate lifespan (Entchev et al., 2015). tph1() and daf7() mutants show respective increases and decreases in food information (Figure 1C), consistent with our prior decoding analysis. tph1() mutants also show a modest decrease in the fraction of redundant information (Figure 1E), suggesting that the added information is more efficiently but less robustly encoded.
Remarkably, changes in the expression distributions of the daf7 and tph1 reporters in daf7() mutants shift the encoding strategy of ASI, ADF, and NSM from redundancy to synergy (Figure 1C–D), such that $\sim 40\%$ of the total information in the circuit is now encoded synergistically (Figure 1E). This effect is not due to the loss of ASI function in daf7() mutants, as we observed the same shift to synergy when only tph1() expressing neurons are analyzed (Figure 1F–H), indicating that crosstalk between daf7 and tph1 as well as daf7 autoregulation control the coding strategy adopted by the circuit. Importantly, the coding strategy shift is daf7specific, as disruption of tph1 does not result in a similar phenotype (Figure 1C). In the tph1(); daf7() double mutant, cross and selfregulation are abolished, and ASI, ADF, and NSM neurons approach the independence regime ($R=0$) (Figure 1C–E), confirming the idea that redundancy and synergy arise from the communication between neurons via daf7 and tph1.
The same informationtheoretic analysis can be applied to quantify more directly the contribution of daf7 and tph1 to the foodresponsiveness of the physiological output. The lifespan response to food abundance consists of $\sim 0.6$ bits of information in wildtype animals, and approximately 80% of this food information is lost in the tph1(); daf7() double mutant (Figure 1I), strengthening our previous assertion that the majority of the food information encoded in the lifespan response is mediated by tph1 and daf7. While other genetic pathways may also play important roles, this central role of tph1 and daf7 suggests that their coding features weigh heavily on the physiological outcome.
Multicellular coding strategies rely on response correlations between cells (Schneidman et al., 2003). Specifically, redundancy can be dissected into two components: the signal correlation, which reflects correlated average responses (Figure 2A) and increases redundancy; and the noise correlation, which captures cofluctuations among different cells under fixed food levels (Figure 2B–C) and promotes synergy (Schneidman et al., 2003) (Appendix). As opposed to the wildtype animals, where the negligible value of noise correlation leads to redundancy (Figure 2D–E), all mutants display a general increase of noise correlations. tph1() animals retain redundancy by compensating this effect with an increase of signal correlation; however, this balance shifts in the daf7() mutant due to the dramatic reduction of signal correlation (Figure 2F), bringing the system to the synergistic regime (Figure 1D). The tph1(); daf7() double mutant has nearly equal signal and noise correlations which generate independent encoding.

Figure 2—source data 1
 https://doi.org/10.7554/eLife.24040.014
Redundancy and synergy is strongly affected by noise and correlation among neurons. To characterize their effects, we rescaled noise and correlations in the original response distributions of daf7 and tph1 over a biologically relevant range (Figure 3, Appendix). In wildtype animals, redundancy is highly sensitive to noise, and weakly sensitive to correlation, providing a rationale for daf7 in noise reduction (Entchev et al., 2015). tph1() mutants displayed increased sensitivity to both noise and correlations. Redundancy in daf7() mutants was more sensitive to correlation than noise, a reversal of the wildtype situation. tph1(); daf7() double mutants were less sensitive to noise and correlations than either single mutant. These results suggest that the sensitivity of redundancy to noise is controlled by daf7, while robustness to correlation is maintained by both daf7 and tph1.
Redundancy or synergy in daf7 and tph1 expressing neurons serves as one constraint but does not necessarily lead to the same coding strategy in their targets. The coding strategy used by these targets will depend on their connectivity to ASI, ADF, and NSM, as well as their noise, correlation, and dynamic range. Since little is known about the immediate targets of TGF$\beta $ and serotonin signaling in relation to the food response in C. elegans, we considered a minimal model of three ideal sensors detecting an input and transmitting to a target that integrates linearly their signals (Figure 4A, Appendix). This simple model shows that decreasing signaltonoise ratio favors synergy (Figure 4B, Appendix), in agreement with the observation that daf7() mutants show reduced signaltonoise, and adopt synergistic encoding (Figure 1D–F). This model also explains the decrease in synergy in tph1(); daf7() double mutants compared to daf7() single mutants (Figure 1D–F): loss of tph1 increases signal separation (Entchev et al., 2015), which increases signaltonoise, thus reducing synergy. Thus, that signaltonoise ratios can contribute significantly to the coding strategy.
Our model also illustrates the advantages of redundancy in the case of linear integration. Redundant strategies increase the minimum information transmitted to a downstream target when compared to a synergistic encoding (Figure 4C). Additionally, redundant encoding not only allows higher information transmission, but can also be accommodated by a broader set of signaling parameters (Figure 4D), avoiding the need to fine tune biological properties. When considering lifespan as the downstream target, our model suggests that lifespan responsiveness to food should decrease in daf7() mutants, because wildtype animals employ redundancy, whereas daf7() mutants employ a synergistic encoding. Indeed, we find that the ability to accurately discriminate between different food inputs based on lifespan is degraded in daf7() mutants (Figure 1I) (Entchev et al., 2015).
By extending the analysis of our previous work, we have found that the ADF, NSM and ASI neurons employ a redundant strategy to encode food information. Critically, this redundant encoding strategy is controlled by daf7 TGF$\beta $ and modified by tph1 tryptophan hydroxylase; this is a novel effect of neuromodulators on circuit function. In particular, we revealed two roles for daf7: as an encoder of food information, and as a regulator of redundancy via regulation of tph1. In principle, redundancy and synergy could be specified by many different biological mechanisms, with obvious candidates being developmental changes in sensor types or numbers in a neural circuit. These mechanisms are ruled out in daf7() and tph1() animals, as the mutations do not affect the development of the ASI, ADF, and NSM neurons, which remain foodresponsive. Instead, we show that daf7 and tph1 influence information processing via effects on the signal and noise properties of these sensory neurons, and on their correlations, representing additional roles for these genes in controlling information encoding. The discovery of other genes that regulate the signaltonoise ratio will likely provide further insights into genetic regulatory mechanisms that modulate neural coding.
Computational methods
Minimization and quantification of experimental noise
Request a detailed protocolInformation theory relies on accurate estimates of response distributions, requiring the minimisation of experimental variability. We took several steps to achieve this. First, we only considered animals oriented in a dorsoventral position. The microfluidic chip was constructed to bias animals towards this correct orientation, the orientation was checked during automated cell identification and verified manually, ensuring that only image stacks with animals in dorsoventral orientations were used in the analysis. Second, we used direct imaging of transcriptional fusions to fluorescent protein reporters integrated in single copy. This approach ensures that biological variance in promoter activity is not artificially washed out by averaging in conventional highcopy reporters that are more traditionally used to generate C. elegans transgenics. Using fluorescent reporters also eliminates experimental noise associated with antibody staining due to variability in fixation, in permeabilizing the C. elegans cuticle, and in signal amplification from secondary antibodies. Third, we minimized bleaching by using a combination of low excitation from an LED light source, and rapid image acquisition using a Piezo Z stage (Prior Scientific) that precisely moves the sample in the Z axis at high speed.
In addition, we used simultaneous quantification of mCherry and Venus/YFP driven by the same promoter to estimate our experimental noise (Figure 1—figure supplement 2). We generated animals with Pdaf7::mCherry and Pdaf7::Venus reporters integrated at single copy in precise genomic locations on LG I and LG II, respectively (Figure 1—figure supplement 2A). These animals were shifted to four different food levels and imaged 1 day after the food shift. This experimental measurement incorporates experimental noise associated with different fluorescent proteins (mCherry and Venus) and different chromosomal locations for reporters, as well as other methodological noise. We found that the two measurements were in good agreement ($R\sim 0.83$, Figure 1—figure supplement 2B). Dissecting the variance in these measurements showed that 30% ($1{R}^{2}$) of the observed variability in these measurements was due to variability between the mCherry and Venus readouts. We note that this variability includes intrinsic noise as the reporters are on different chromosomes; the actual experimental variability would therefore be lower, since intrinsic noise is nonzero.
Computational analysis
Request a detailed protocolThe computational analysis of all the data was performed using custommade C++ programs and builtin implementations of standard multivariate analysis algorithms in R (R Core Team, 2016). C++ programs are available through GitHub repositories (https://github.com/giovannidiana/Information, https://github.com/giovannidiana/KDE and https://github.com/giovannidiana/ModelRS). Mathematical details of these procedures and the results are discussed in the Appendix.
Appendix
Supplementary computational methods
To uncover the information processing features of the daf7/tph1 genetic circuit embedded in the ASI, ADF and NSM neurons, we performed an informationtheoretic analysis of the gene expression responses of daf7 in the ASI neuron, and tph1 in the ADF and NSM neurons. In the main text we introduced the mutual information (MI) as a measure of the correlation between food level and gene expression. In this section we discuss in greater detail the properties of this quantity and the procedure used to estimate MI from our gene expression data. The type of encoding system that we are interested in maps an input $F$ (food level) taking ${N}_{F}=6$ distinct values $\{{f}_{1},\mathrm{\cdots},{f}_{{N}_{F}}\}$ onto three continuous variables denoted by the vector $G=\{{G}_{ADF},{G}_{ASI},{G}_{NSM}\}$ (gene expression in the three neurons).
Mutual information
The multivariate gene expression response under a specific food condition is given by the set of conditional probabilities $P\left(G\rightF={f}_{k})$, ($k=1,\mathrm{\dots},{N}_{F}$). To characterize the information transmission of a communication system we also need to specify the probabilities $P\left(F\right)$ with which the ADFASINSM encoder is exposed to each food condition. Given the input probabilities $P\left(F\right)$ we can compute averages across food level, in particular the marginal probabilities of gene expression
From input and response distributions we can build up three information entropies. First, the joint information entropy of both food and gene expression is defined as (Cover and Thomas, 2006)
and it measures the variability of input and output. Second, we can quantify the variability of the gene expression response to food by the conditional entropy
Third, the entropy of the marginal distributions in Equation (3)
measures the variability of the average response. The mutual information defined as
is always positive due to the logsum inequality and it can be expressed as the difference between the gene expression entropy and the conditional entropy with respect to food level, that is
which yields to the standard interpretation of $MI$ as the amount of information entropy shared between stochastic variables.
As mentioned in the main text and Figure 1A, the mutual information is strongly affected by the signaltonoise ratio ($SNR$). For univariate distributions, we use the definition
where $\mathrm{var}\left(G\rightF)$ is the variance under the condition $F$ and ${\u27e8\cdot \u27e9}_{F}$ denotes the average across all conditions.
MI can be decomposed as
where the components ${D}_{k}^{\left(joint\right)}$ are defined as
and represent the relative entropy between conditional and average response.
Our estimates of mutual information provide a lower bound of the true information encoded by tph1 and daf7 due to noise inherent in all experiments.
Channel capacity
A common question in biology is to understand how phenotypic changes are related to the environmental input. To address this question it is natural to design experiments where relevant input variables are controlled. These types of experiments provide a good sampling of the responses, but the frequencies of environmental conditions at which biological systems are exposed in the wild are not always known. On the other hand, the level of information encoded about the environment depends on the input distribution. A common procedure to infer the input distribution is to assume that the set of gene expression responses is designed to maximize the information stored (Tkačik and Walczak, 2011; Selimkhanov et al., 2014; Uda et al., 2013). With this assumption we can obtain the food distribution by maximising the mutual information between food and gene expression. The maximal MI achievable given the set of conditional responses is known as channel capacity
and it is an intrinsic property of the encoding system.
An important aspect of MI is that all the relative entropies ${D}_{k}^{\left(joint\right)}$ in the decomposition (10) become identical under the optimality condition, which is easy to prove by maximizing the action
where $\lambda $ is a Lagrange multiplier to assure the normalization of the input distribution. By considering the derivative over $P\left({f}_{k}\right)$ we get
which implies that at the maximum, all the ${D}_{k}^{\left(joint\right)}$ are equal to $1\lambda $. Therefore, since the channel capacity is defined as an average of ${D}_{k}^{\left(joint\right)}$, ${D}_{k}^{\left(joint\right)}=C$ for all $k$. This property implies that the optimal input distribution obtained by maximizing MI is such that all the conditional responses are equally distant from their average, when relative entropy is used as a measure of distance between probability distributions.
In Figure 1—figure supplement 5A we compare the channel capacity obtained from tph1/daf7 expression in ADF, ASI and NSM neurons (dotted line) with the components ${D}_{k}^{\left(neuron\right)}$ of the mutual information between gene expression and food abundance in each neuron (ADF, ASI, NSM) for all genetic backgrounds. For this comparison we first obtained the channel capacity and the optimal input distribution from the threedimensional data for each genotype (Figure 1—figure supplement 5C) and then we used this optimal distribution to calculate the mutual information of each neuron. The components of the mutual information for individual neurons are obtained from Equation (11) by using the corresponding marginal distribution. Unlike the components ${D}_{k}^{\left(joint\right)}$ of the maximized joint mutual information, the ${D}_{k}^{\left(neuron\right)}$ are not constant over the food level, reflecting the fact that single neurons are optimized for different input distributions (Figure 1—figure supplement 3).
The optimal input frequencies (Figure 1—figure supplement 5C) reveal that wildtype animals encode the most information when they are most likely to encounter the highest, an intermediate, and the lowest food levels. This result implies that wildtype animals are best at detecting these food levels, compatible with the boom and bust lifestyle of C. elegans in the wild (Félix and Braendle, 2010). This optimal is altered in daf7() mutants (Figure 1—figure supplement 5C), indicating that it is genetically controlled. By maximizing the mutual information between individual neurons and food conditions we find that each neuron is specialized to sense different food levels (Figure 1—figure supplement 3), which broadens their combined range of detectable food levels. For example, tph1 expression in ADF in wildtype animals is best at detecting the food extremes (Figure 1—figure supplement 3A). At these extreme food levels, ADF carries more information than at other food levels. Thus specialization among food sensing neurons ultimately leads to fooddependent heterogeneity in coding (Figure 1—figure supplement 5A).
We note that the switch from redundancy in wildtype to synergy in daf7() mutants still occurs when we use the wildtype optimal input frequency for calculating redundancy values for daf7() mutants. Thus, our conclusions are not sensitive to the choice of using channel capacity and the corresponding optimal input distribution for each genotype. All the estimates of channel capacity in this work were done by using the standard ArimotoBlahut algorithm (Arimoto, 1972; Blahut, 1972).
Redundancy and synergy
In the main text we introduce redundancy $R$ as the difference (Schneidman et al., 2003)
and synergy as the negative of the redundancy. By following the work of Schneidman et al. the redundancy can be written as the difference between signal and noise correlation defined as
where ${I}^{\left(s\right)}$ is the ‘shuffle’ information defined as
and corresponds to a modified version of the mutual information between gene expression and food level where the joint distribution is replaced by the product of the marginal densities ${P}^{\left(s\right)}$,
By using the definition of the mutual information per neuron we can rewrite the signal correlation in the form of a relative entropy
which shows that $SC$ is a nonnegative quantity. As opposite to the signal correlation, which can only increase the level of redundancy, noise correlation can be positive or negative. Depending on the sign of $R$ in Equation (15) the system operates in a redundant ($SC>NC$) or synergistic ($SC<NC$) regime.
These informationtheoretic measures of correlation reveal changes in different genotypes that contribute to shifts in coding strategy and capture different features of the interaction between the neurons. Consider for instance the case of independent encoders where the probability distribution of the joint response is factorized into the product of the responses of each neuron
In this case the noise correlation vanishes identically, however, signal correlation can be nonzero due to the correlation induced by the stimulus, thus we obtain the intuitive result that the level of redundancy in a system of independent encoders is always nonnegative. Therefore, the synergistic encoding that we observe in the daf7() mutant is caused by the change in the interaction network of ADF, ASI and NSM neurons. In the wildtype this network is tuned to guarantee a robust encoding of food abundance. When daf7 is knockedout, the sign of $R$ in Equation (15) changes, which namely corresponds to a switch from redundancy to synergy.
Analogously to the mutual information, also redundancy can be decomposed as
where we defined the food components ${D}_{k}^{\left(red\right)}$ as
In Figure 1—figure supplement 5B we show the redundancy components at each food level across all genotypes. We observe that the quality of the encoding (synergistic or redundant) varies under different food conditions. In particular, both wildtype and tph1() mutant tend to adopt a synergistic behaviour under non optimal food conditions (see Figure 1—figure supplement 3A–B for comparison with input distributions) whereas daf7() mutant and tph1(); daf7() double mutant are always synergistic. The input distribution obtained by maximizing $MI({G}_{ADF},{G}_{ASI},{G}_{NSM};F)$ was used as a reference food distribution for all the genotypes analyzed in this work. tph1 and daf7 promoter activity was available also for mutant strains because the reporters were separate from the endogenous genes. To confirm that the synergistic character of the encoding in the daf7() mutant is not an artifact of including ASI (where daf7 is expressed) in the estimation of the redundancy, we performed the same analysis by using only ADF and NSM readouts. As a result, by comparing wildtype and daf7() mutant, we obtained the same qualitative switch from redundant to synergistic encoding as obtained from the inclusion of all neurons (Figure 1F–H).
Kernel density estimation
Information entropies, and thus mutual information, are functionals of the probability distribution of the readouts. To quantify the conditional distributions $P\left(G\rightF)$ we used Kernel Density Estimation (KDE) (Scott, 1992), which provides a mathematical framework to estimate distributions of continuous variables.
Compared to the standard methodology of frequency histograms to estimate distributions, this technique does not require bin size selection. In the KDE approach, the probability density is estimated by the sum of reference distributions (kernel) centered at the observed values, thus for any expression vector $g$ we have the estimated density $\hat{f}(g)$ reads
where the kernel ${K}_{H}$ is a multivariate Gaussian distribution, the ‘bandwidth’ $H$ corresponds to its variance matrix and the sum is over all the measured expressions ${\left\{{G}_{i}\right\}}_{i=1}^{n}$.
An accurate estimation of the density relies on the choice of the bandwidth, which can be constant across the support of the probability or adapted to the local density. The Mean Squared Error (MSE)
and its integral (MISE) are commonly minimized to find the appropriate bandwidth. Selector algorithms differ in the tradeoff between bias and variance of the estimator.
To check the robustness of our calculation, we compared the results obtained by using different fixed bandwidth selector algorithms (Figure 1—figure supplement 4). In particular, we used the plugin method (Chacón and Duong, 2010), least squares crossvalidation (Bowman, 1984) and smoothing cross validation (Jones et al., 1991), all of which provide a uniform bandwidth. The general, fixed bandwidth estimators tends to oversmooth the main part of the distribution and undersmooth the tails. To confirm that this effect did not introduce artificial biases we also used the ‘baloon’ ($k$nearest neighbours) estimator (Loftsgaarden and Quesenberry, 1965), where the probability distribution is proportional to the local density of observations (Figure 1—figure supplement 4).
Once we obtained the conditional response distributions, averages over expression levels as in Equation (7) were computed by evaluating $P\left(G\rightF)$ on a threedimensional grid (a different approach would be to resample from the obtained distribution (Krishnaswamy et al., 2014). By testing different grid resolutions we found that a grid of size ${30}^{3}$ was sufficient to guarantee the convergence of averages. The uncertainty in the estimation of channel capacity was obtained by calculating the variance associated with sampling the 80% of the data. As shown in Figure 1—figure supplement 4, the estimates of both channel capacity and redundancy/synergy are robust to KDE algorithm in all genetic backgrounds.
Sample size bias
A well known issue in the estimation of channel capacity is the bias due to sample size. The general jackknife procedure to remove this effect involves expanding the channel capacity in inverse powers of sample size (Cheong et al., 2011; Selimkhanov et al., 2014),
and obtaining the unbiased term by a linear fit of the channel capacity calculated using increasing fraction of the data. By applying this procedure, we found a very small samplesize correction to channel capacity in all genetic backgrounds (Figure 1—figure supplement 4A–B). The same analysis applied to the redundancy/synergy (Figure 1—figure supplement 4C–D), showed that our conclusions are independent on the sample size. All our linear fits of channel capacity and redundancy/synergy (Figure 1—figure supplement 4E) from 60% to 100% of the data were above the 95% of confidence level, indicating that our data is far from the undersampled regime.
Covariance sensitivity analysis
To explore how channel capacity and redundancy depend on linear correlation and noise among ADF, ASI and NSM neurons requires a way to scale these two properties in silico from the baselines obtained in each genotype in experimental measurements. To do so, we first approximated the gene expression densities as multivariate normal distributions. This approximation captures most of the global features of our threedimensional responses and allows us to control noise and correlations in terms of covariance matrices. The Gaussian assumption was also used in our previous decoding analysis (Entchev et al., 2015). The agreement between our present study and the decoding analysis shows indirectly that the Gaussian approximation can be used here for informationtheoretic purposes. We used the maximumlikelihood estimates of the covariance matrices $C$ for each genotype as a reference and then we transformed each entry of the covariance matrix according to the rule
The transformation above rescales all the standard deviations of the responses by a factor $\alpha $ and the Pearson’s correlation index for all pairs of neurons by a factor $\beta $. Thus we studied the sensitivity of informationtheoretic variables to noise and correlation by varying $\alpha $ and $\beta $ over a biologically relevant range.
In the main text we presented the sensitivity analysis of redundancy, in Figure 3A and Figure 3—figure supplement 1A–C we show the colorcoded contour maps of channel capacity, signal and noise correlation obtained by varying the parameters $\alpha $ and $\beta $ from 0.5 to 2 independently. We checked numerically the positivity of the covariance matrix for all pairs of $\alpha $ and $\beta $. The major factor that controls information capacity in wildtype is noise, which provides a rationale for the noise regulation by daf7 revealed in our previous study (Entchev et al., 2015). Scaling the linear correlation has a more pronounced effect in all the mutants and especially in the daf7() mutant. This is due to the synergistic encoding in daf7() mutants  since interactions between neurons are a crucial to a synergistic strategy, the system becomes much more sensitive to linear correlations. This effect is particularly evident in daf7() mutants, where the signal correlation is almost unchanged under noise rescaling (Figure 3—figure supplement 1C), making noise correlation the more prominent contributing factor.
Gaussian model
We can explore the consequences of redundant or synergistic strategies by modeling how the information encoded by ADF, ASI and NSM neuron is read by an ideal output which conveys the information from the sensory neurons. The essential features of the ADFASINSM system are captured by using the model depicted in Figure 4—figure supplement 1A–B. Here, the information about a binary input $B$ is encoded by three sensors ${S}_{1}$, ${S}_{2}$ and ${S}_{3}$ whose joint response is a multivariate normal distribution with mean vector ${\mu}_{b}$
where the parameter $a$ is associated with the dynamic range of the response. The covariance matrix ${C}_{ij}\equiv \mathrm{cov}({S}_{i},{S}_{j})$ associated to the joint distribution is assumed to be stimulusindependent. This simplification is consistent with the observation that variances and correlations between neurons do not change considerably across food levels. The covariance matrix was parametrized as
where ${\rho}_{ij}$ are the correlation coefficients between ${S}_{i}$ and ${S}_{j}$ such that $detC>0$.
The information about the binary input encoded by the three Gaussian sensors is then integrated linearly by the output variable
where $\alpha ,\beta ,\gamma >0$ and $\alpha +\beta +\gamma =1$. This choice implies that the output is also normally distributed with mean $\pm a$ dependently on the value of the input, whereas its variance reads
In our setting we assume the two states of the input $b\in \{0,1\}$ to be equally probable, leading to an information entropy of 1 bit. The information encoded by the sensors about the input, $I=MI({S}_{1},{S}_{2},{S}_{3};B)$, is upper bounded by the input entropy, moreover the input information encoded by each component, ${I}_{k}=MI({S}_{k};B)$, is always smaller than the joint information. We can combine these constraints into the inequality
Furthermore, since the output is a function of the sensor responses, the mutual information between input and output ${I}_{o}=MI(\mathrm{output};B)$ is bounded by the information encoded by the three sensors
In order to understand the consequences of synergy and redundancy from the perspective of the output node in the network which receives the input information from the three sensors, we explored the parametric space of the model and calculated the information encoded by the sensors and by the output. By using the variance of the first sensor as a reference scale we set ${\sigma}_{1}=1$ and sampled the eight parameters left uniformly within the range
The range above was selected based on the following considerations:
All variances have a lower bound (set to 0.1) to avoid singular regimes where the 3D normal distribution becomes too narrow around the mean.
From Equations (34,35) $\sigma}_{1}\text{}{\sigma}_{2}\text{}\text{}{\sigma}_{3$, which implies $I}_{1}\text{}\text{}{I}_{2}\text{}\text{}{I}_{3$.
The upper bound of 0.7 on the absolute correlation coefficients ${\rho}_{ij}$ was used to keep the correlations within a biologically relevant range. Correlations between ADF, ASI and NSM are lower than 0.5 in all food conditions and genetic backgrounds.
The conditions in Equations (34–36) do not guarantee the covariance matrix in Equation (29) to be positive definite, therefore in our sampling algorithm we rejected all parameter sets with $\mathrm{det}\left(\mathrm{cov}\right)<0$.
In our sampling we choose the value of $a$ to be lower than $2{\sigma}_{3}$. Larger values of $a$ generate extreme regimes where ${I}_{3}$ is approximately one bit, and the inequality for the joint information ${I}_{3}\le I\le 1$ implies a positive redundancy
In this condition the output information is very sensitive to the value of $\gamma =1\alpha \beta $. For $\gamma \approx 1$, the output is only receiving input from ${S}_{3}$, leading to an efficient transmission of 1 bit of information. Lower values of $\gamma $ lead to a decrease of the transmitted information due to the noisier contribution of ${S}_{1}$ and ${S}_{2}$ to the output.
In Figure 4 and Figure 4—figure supplement 1 we show the calculation of informationtheoretic quantities from a sample of $\sim 500000$ parametric sets. Red and blue populations correspond respectively to redundant and synergistic configurations. The majority of the sampled configurations (65%) displays a positive redundancy. As discussed in the main text, the minimum value of the output information increases proportionally to the level of redundancy. This feature matches the intuitive view that redundant system allow to transmit infomation more reliably. The redundancy value is lower bounded by the negative of the total information encoded by the sensors and upper bounded by two bits (Figure 4—figure supplement 1C), due to the inequalities
Synergistic regimes occupy the region of low signal correlation and positive noise correlation (Figure 4—figure supplement 1D–F) and are generally characterized by low values of the information carried by single sensors (Figure 4—figure supplement 1E).
In Figure 4—figure supplement 1G, we show the distribution of redundant/synergistic regimes with respect to the parameters $a$ and ${\sigma}_{3}$, which represent dynamic range and noise in the model. The ratio between these two parameters quantifies the signaltonoise ratio of the system
In the absence of extra constraints, redundant configurations are permitted for any value of the signaltonoise ratio, whereas the population of synergistic regimes is depleted for high values of $SNR$ (Figure 4—figure supplement 1H). When we require a nonzero lower bound to the information encoded by the sensors, $I\text{}\text{}{I}_{\mathrm{m}\mathrm{i}\mathrm{n}}$, we see the appearance of a critical value for $SNR={s}^{*}\left({I}_{0}\right)$ which separates two regions (see Figure 4—figure supplement 1H, right panel): a synergydominated region, for $SNR\text{}\text{}{s}^{\ast}({I}_{0})$, and a mixed region where both coding strategies are permitted $SNR\text{}\text{}{s}^{\ast}({I}_{0})$. The critical $SNR$ value depends on threshold applied to the sensor information, in particular, ${s}^{*}\left({I}_{0}\right)$ increases for increasing threshold ${I}_{0}$.
This observation can be used to predict how changes in signaltonoise ratio affect coding strategy. Consider a system operating redundantly at high $SNR$. Our model shows that independently of the details of the system, if we apply a perturbation to the system which reduces the $SNR$ below ${s}^{*}\left({I}_{0}\right)$, then in order to carry at least ${I}_{0}$ bits of information the system will necessarily adopt a synergistic strategy. Remarkably, this feature of the model is in perfect agreement with the switch from redundancy to synergy observed in the daf7() mutant with respect to the wildtype animal.
Reduction of $SNR$ accompanied by a sufficient level of information encoded is always associated to a switch to synergy. This behaviour is easy to explain. When $a\ll {\sigma}_{3}$ the most informative sensor ${S}_{3}$ stores a very small amount of information due to the small $SNR$ however the joint information can still reach one bit by increasing the eccentricity of the distributions, i.e. by increasing the linear correlations between sensors. This has the clear consequence of increasing the noise correlation, therefore shifting redundancy to negative values.
References

An algorithm for computing the capacity of arbitrary discrete memoryless channelsIEEE Transactions on Information Theory 18:14–20.https://doi.org/10.1109/TIT.1972.1054753

Neural correlations, population coding and computationNature Reviews Neuroscience 7:358–366.https://doi.org/10.1038/nrn1888

Computation of channel capacity and ratedistortion functionsIEEE Transactions on Information Theory 18:460–473.https://doi.org/10.1109/TIT.1972.1054855

Environmental sensing, information transfer, and cellular decisionmakingCurrent Opinion in Biotechnology 28:149–155.https://doi.org/10.1016/j.copbio.2014.04.010

Synergy in a neural codeNeural Computation 12:1531–1552.https://doi.org/10.1162/089976600300015259

The natural history of Caenorhabditis elegansCurrent Biology 20:R965–R969.https://doi.org/10.1016/j.cub.2010.09.050

The sensory system: More than just a window to the external worldCommunicative & Integrative Biology 8:e1017159.https://doi.org/10.1080/19420889.2015.1017159

A simple root n bandwidth selectorThe Annals of Statistics 19:1919–1932.https://doi.org/10.1214/aos/1176348378

Cellular noise and information transmissionCurrent Opinion in Biotechnology 28:156–164.https://doi.org/10.1016/j.copbio.2014.05.002

A nonparametric estimate of a multivariate density functionThe Annals of Mathematical Statistics 36:1049–1051.https://doi.org/10.1214/aoms/1177700079

SoftwareR: a language and environment for statistical computingR Foundation for Statistical Computing, Vienna, Austria.

Synergy, redundancy, and independence in population codesJournal of Neuroscience 23:11539–11553.

Synergy from silence in a combinatorial neural codeJournal of Neuroscience 31:15732–15741.https://doi.org/10.1523/JNEUROSCI.030109.2011

A mathematical theory of communicationBell System Technical Journal 27:379–423.https://doi.org/10.1002/j.15387305.1948.tb01338.x

Information transmission in genetic regulatory networks: a reviewJournal of Physics: Condensed Matter 23:153102.https://doi.org/10.1088/09538984/23/15/153102
Decision letter

Oliver HobertReviewing Editor; Howard Hughes Medical Institute, Columbia University, United States
In the interests of transparency, eLife includes the editorial decision letter and accompanying author responses. A lightly edited version of the letter sent to the authors after peer review is shown, indicating the most substantive concerns; minor comments are not usually included.
[Editors’ note: minor issues and corrections have not been included, so there is not an accompanying Author response.]
Thank you for submitting your article "Genetic Control of Encoding Strategy in a Foodsensing Neural Circuit" for consideration as a Research Advance by eLife. Your article has been reviewed by three peer reviewers, and the evaluation has been overseen by a Reviewing Editor (Oliver Hobert) and a Senior Editor. The reviewers have opted to remain anonymous.
The reviewers have discussed the reviews with one another and are in agreement that this work qualifies as an interesting Research Advance to your original eLife paper.
The only comment we would like you address is extremely minor in nature: Figure 1—figure supplement 3: panels AD are shown, but the legend refers to AE.
https://doi.org/10.7554/eLife.24040.020Article and author information
Author details
Funding
National Science Foundation (0946809 GRFP)
 Mei Zhan
National Institutes of Health (R01AG035317)
 Hang Lu
National Science Foundation (0954578)
 Hang Lu
National Institutes of Health (R01GM088333)
 Hang Lu
Wellcome (087146)
 QueeLim Ch'ng
Biotechnology and Biological Sciences Research Council (BB/H020500/1)
 QueeLim Ch'ng
European Commission (NeuroAge 242666)
 QueeLim Ch'ng
Biotechnology and Biological Sciences Research Council (BB/M00757X/1)
 QueeLim Ch'ng
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
Acknowledgements
We thank the Bargmann and Horvitz labs for reagents. Some strains were provided by the CGC, which is funded by NIH Office of Research Infrastructure Programs (P40 OD010440). We also thank R Endres, J Clarke, S Oliferenko, P Gordon, E Makeyev, O Marin, and D Passaro for comments on the manuscript; and B Handley, K GersBarlag, O Leyshon, and H Tunbak for technical assistance. This research was supported by the Wellcome Trust (Project Grant 087146 to QC), BBSRC (BB/H020500/1 and BB/M00757X/1 to QC), European Research Council (NeuroAge 242666 to QC), US National Institutes of Health (R01AG035317 and R01GM088333 to HL), and US National Science Foundation (0954578 to HL, 0946809 GRFP to MZ).
Reviewing Editor
 Oliver Hobert, Howard Hughes Medical Institute, Columbia University, United States
Publication history
 Received: December 8, 2016
 Accepted: January 16, 2017
 Version of Record published: February 7, 2017 (version 1)
Copyright
© 2017, Diana et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics

 1,502
 Page views

 244
 Downloads

 5
 Citations
Article citation count generated by polling the highest count across the following sources: PubMed Central, Crossref, Scopus.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading

 Computational and Systems Biology
MicroRNAs (miR), as important epigenetic control factors, reportedly regulate wound repair. However, our insufficient knowledge of clinically relevant miRs hinders their potential therapeutic use. For this, we performed paired small and long RNAsequencing and integrative omics analysis in human tissue samples, including matched skin and acute wounds collected at each healing stage and chronic nonhealing venous ulcers (VUs). On the basis of the findings, we developed a compendium (https://www.xulandenlab.com/humanwoundsmirnamrna), which will be an open, comprehensive resource to broadly aid wound healing research. With this first clinical, woundcentric resource of miRs and mRNAs, we identified 17 pathologically relevant miRs that exhibited abnormal VU expression and displayed their targets enriched explicitly in the VU gene signature. Intermeshing regulatory networks controlled by these miRs revealed their high cooperativity in contributing to chronic wound pathology characterized by persistent inflammation and proliferative phase initiation failure. Furthermore, we demonstrated that miR34a, miR424, and miR516, upregulated in VU, cooperatively suppressed keratinocyte migration and growth while promoting inflammatory response. By combining miR expression patterns with their specific target gene expression context, we identified miRs highly relevant to VU pathology. Our study opens the possibility of developing innovative wound treatment that targets pathologically relevant cooperating miRs to attain higher therapeutic efficacy and specificity.

 Computational and Systems Biology
 Neuroscience
The maintenance of items in working memory (WM) relies on a widespread network of cortical areas and hippocampus where synchronization between electrophysiological recordings reflects functional coupling. We investigated the direction of information flow between auditory cortex and hippocampus while participants heard and then mentally replayed strings of letters in WM by activating their phonological loop. We recorded local field potentials from the hippocampus, reconstructed beamforming sources of scalp EEG, and – additionally in four participants – recorded from subdural cortical electrodes. When analyzing Granger causality, the information flow was from auditory cortex to hippocampus with a peak in the [4 8] Hz range while participants heard the letters. This flow was subsequently reversed during maintenance while participants maintained the letters in memory. The functional interaction between hippocampus and the cortex and the reversal of information flow provide a physiological basis for the encoding of memory items and their active replay during maintenance.