Adaptation of olfactory receptor abundances for efficient coding
Abstract
Olfactory receptor usage is highly heterogeneous, with some receptor types being orders of magnitude more abundant than others. We propose an explanation for this striking fact: the receptor distribution is tuned to maximally represent information about the olfactory environment in a regime of efficient coding that is sensitive to the global context of correlated sensor responses. This model predicts that in mammals, where olfactory sensory neurons are replaced regularly, receptor abundances should continuously adapt to odor statistics. Experimentally, increased exposure to odorants leads variously, but reproducibly, to increased, decreased, or unchanged abundances of different activated receptors. We demonstrate that this diversity of effects is required for efficient coding when sensors are broadly correlated, and provide an algorithm for predicting which olfactory receptors should increase or decrease in abundance following specific environmental changes. Finally, we give simple dynamical rules for neural birth and death processes that might underlie this adaptation.
https://doi.org/10.7554/eLife.39279.001eLife digest
A mouse’s nose contains over 10 million receptor neurons divided into about 1,000 different types, which detect airborne chemicals – called odorants – that make up smells. Each odorant activates many different receptor types. And each receptor type responds to many different odorants. To identify a smell, the brain must therefore consider the overall pattern of activation across all receptor types. Individual receptor neurons in the mammalian nose live for about 30 days, before new cells replace them. The entire population of odorant receptor neurons turns over every few weeks, even in adults.
Studies have shown that some types of these receptor neurons are used more often than others, depending on the species, and are therefore much more abundant. Moreover, the usage patterns of different receptor types can also change when individual animals are exposed to different smells. Teşileanu et al. set out to develop a computer model that can explain these observations.
The results revealed that the nose adjusts its odorant receptor neurons to provide the brain with as much information as possible about typical smells in the environment. Because each smell consists of multiple odorants, each odorant is more likely to occur alongside certain others. For example, the odorants that make up the scent of a flower are more likely to occur together than alongside the odorants in diesel. The nose takes advantage of these relationships by adjusting the abundance of the receptor types in line with them. Teşileanu et al. show that exposure to odorants leads to reproducible increases or decreases in different receptor types, depending on what would provide the brain with most information.
The number of odorant receptor neurons in the human nose decreases with time. The current findings could help scientists understand how these changes affect our sense of smell as we age. This will require collaboration between experimental and theoretical scientists to measure the odors typical of our environments, and work out how our odorant receptor neurons detect them.
https://doi.org/10.7554/eLife.39279.002Introduction
The sensory periphery acts as a gateway between the outside world and the brain, shaping what an organism can learn about its environment. This gateway has a limited capacity (Barlow, 1961), restricting the amount of information that can be extracted to support behavior. On the other hand, signals in the natural world typically contain many correlations that limit the unique information that is actually present in different signals. The efficientcoding hypothesis, a key normative theory of neural circuit organization, puts these two facts together, suggesting that the brain mitigates the issue of limited sensory capacity by eliminating redundancies implicit in the correlated structure of natural stimuli (Barlow, 1961; van Hateren, 1992a). This idea has led to elegant explanations of functional and circuit structure in the early visual and auditory systems (see, e.g. Laughlin, 1981; Atick and Redlich, 1990; Van Hateren, 1993; Olshausen and Field, 1996; Simoncelli and Olshausen, 2001; Fairhall et al., 2001; Lewicki, 2002; Ratliff et al., 2010; Garrigan et al., 2010; Tkacik et al., 2010; Hermundstad et al., 2014; Palmer et al., 2015; Salisbury and Palmer, 2016). These classic studies lacked a way to test causality by predicting how changes in the environment lead to adaptive changes in circuit composition or architecture. We propose that the olfactory system provides an avenue for such a causal test because receptor neuron populations in the mammalian nasal epithelium are regularly replaced, leading to the possibility that their abundances might adapt efficiently to the statistics of the environment.
The olfactory epithelium in mammals and the antennae in insects are populated by large numbers of olfactory sensory neurons (OSNs), each of which expresses a single kind of olfactory receptor. Each type of receptor binds to many different odorants, and each odorant activates many different receptors, leading to a complex encoding of olfactory scenes (Malnic et al., 1999). Olfactory receptors form the largest known gene family in mammalian genomes, with hundreds to thousands of members, owing perhaps to the importance that olfaction has for an animal’s fitness (Buck and Axel, 1991; Tan et al., 2015; Chess et al., 1994). Independently evolved large olfactory receptor families can also be found in insects (Missbach et al., 2014). Surprisingly, although animals possess diverse repertoires of olfactory receptors, their expression is actually highly nonuniform, with some receptors occurring much more commonly than others (Rospars and Chambille, 1989; IbarraSoria et al., 2017). In addition, in mammals, the olfactory epithelium experiences neural degeneration and neurogenesis, resulting in replacement of the OSNs every few weeks (Graziadei and Graziadei, 1979). The distribution of receptors resulting from this replacement has been found to have a mysterious dependence on olfactory experience (Schwob et al., 1992; Santoro and Dulac, 2012; Zhao et al., 2013; Dias and Ressler, 2014; Cadiou et al., 2014; IbarraSoria et al., 2017): increased exposure to specific ligands leads reproducibly to more receptors of some types, and no change or fewer receptors of other types.
Here, we show that these puzzling observations are predicted if the receptor distribution in the olfactory epithelium is organized to present a maximally informative picture of the odor environment. Specifically, we propose a model for the quantitative distribution of olfactory sensory neurons by receptor type. The model predicts that in a noisy odor environment: (a) the distribution of receptor types will be highly nonuniform, but reproducible given fixed receptor affinities and odor statistics; and (b) an adapting receptor neuron repertoire should reproducibly reflect changes in the olfactory environment; in a sense it should become what it smells. Precisely such findings are reported in experiments (Schwob et al., 1992; Santoro and Dulac, 2012; Zhao et al., 2013; Dias and Ressler, 2014; Cadiou et al., 2014; IbarraSoria et al., 2017).
In contrast to previous work applying efficientcoding ideas to the olfactory system (Keller and Vosshall, 2007; McBride et al., 2014; Zwicker et al., 2016; Krishnamurthy et al., 2017), here we take the receptor–odorant affinities to be fixed quantities and do not attempt to explain their distribution or their evolution and diversity across species. Instead, we focus on the complementary question of the optimal way in which the olfactory system can use the available receptor genes. This allows us to focus on phenomena that occur on faster timescales, such as the reorganization of the receptor repertoire as a result of neurogenesis in the mammalian epithelium.
Because of the combinatorial nature of the olfactory code (Malnic et al., 1999; Stopfer et al., 2003; Stevens, 2015; Zhang and Sharpee, 2016; Zwicker et al., 2016; Krishnamurthy et al., 2017) receptor neuron responses are highly correlated. In the absence of such correlations, efficient coding predicts that output power will be equalized across all channels if transmission limitations dominate (Srinivasan et al., 1982; Olshausen and Field, 1996; Hermundstad et al., 2014), or that most resources will be devoted to receptors whose responses are most variable if input noise dominates (van Hateren, 1992a; Hermundstad et al., 2014). Here, we show that the optimal solution is very different when the system of sensors is highly correlated: the adaptive change in the abundance of a particular receptor type depends critically on the global context of the correlated responses of all the receptor types in the population—we refer to this as contextdependent adaptation.
Correlations between the responses of olfactory receptor neurons are inevitable not only because the same odorant binds to many different receptors, but also because odors in the environment are typically composed of many different molecules, leading to correlations between the concentrations with which these odorants are encountered. Furthermore, there is no way for neural circuitry to remove these correlations in the sensory epithelium because the candidate lateral inhibition occurs downstream, in the olfactory bulb. As a result of these constraints, for an adapting receptor neuron population, our model predicts that increased activation of a given receptor type may lead to more, fewer or unchanged numbers of the receptor, but that this apparently sporadic effect will actually be reproducible between replicates. This counterintuitive prediction matches experimental observations (Santoro and Dulac, 2012; Zhao et al., 2013; Cadiou et al., 2014; IbarraSoria et al., 2017).
Olfactory response model
In vertebrates, axons from olfactory neurons converge in the olfactory bulb on compact structures called glomeruli, where they form synapses with dendrites of downstream neurons (Hildebrand and Shepherd, 1997); see Figure 1a. To good approximation, each glomerulus receives axons from only one type of OSN, and all OSNs expressing the same receptor type converge onto a small number of glomeruli, on average about two in mice to about 16 in humans (Maresh et al., 2008). Similar architectures can be found in insects (Vosshall et al., 2000).
The anatomy shows that in insects and vertebrates, olfactory information passed to the brain can be summarized by activity in the glomeruli. We treat this activity in a firingrate approximation, which allows us to use available receptor affinity data (Hallem and Carlson, 2006; Saito et al., 2009). This approximation neglects individual spike times, which can contain important information for odor discrimination in mammals and insects (Resulaj and Rinberg, 2015; DasGupta and Waddell, 2008; Wehr and Laurent, 1996; Huston et al., 2015). Given data relating spike timing and odor exposure for different odorants and receptors, we could use the time from respiratory onset to the first elicited spike in each receptor as an indicator of activity in our model. Alternatively, we could use both the timing and the firing rate information together. Such data is not yet available for large panels of odors and receptors, and so we leave the inclusion of timing effects for future work.
A challenge specific to the study of the olfactory system as compared to other senses is the limited knowledge we have of the space of odors. It is difficult to identify common features shared by odorants that activate a given receptor type (Rossiter, 1996; Malnic et al., 1999), while attempts at defining a notion of distance in olfactory space have had only partial success (Snitz et al., 2013), as have attempts to find reduceddimensionality representations of odor space (Zarzo and Stanton, 2006; Koulakov et al., 2011). In this work, we simply model the olfactory environment as a vector $\mathbf{\mathbf{c}}=\{{c}_{1},\mathrm{\dots},{c}_{N}\}$ of concentrations, where ${c}_{i}$ is the concentration of odorant $i$ in the environment (Figure 1a). We note, however, that the formalism we describe here is equally applicable for other parameterizations of odor space: the components ${c}_{i}$ of the environment vector $\mathbf{\mathbf{c}}$ could, for instance, indicate concentrations of entire classes of molecules clustered based on common chemical traits, or they might be abstract coordinates in a lowdimensional representation of olfactory space.
Once a parameterization for the odor environment is chosen, we model the statistics of natural scenes by the joint probability distribution $P({c}_{1},\mathrm{\dots},{c}_{N})$. We are neglecting temporal correlations in olfactory cues because we are focusing on odor identity rather than olfactory search where timing of cues will be especially important. This simplifies our model, and also reduces the number of olfactory scene parameters needed as inputs. Similar static approximations of natural images have been employed powerfully along with the efficient coding hypothesis to explain diverse aspects of early vision (e.g., in Laughlin, 1981; Atick and Redlich, 1990; Olshausen and Field, 1996; van Hateren and van der Schaaf, 1998; Ratliff et al., 2010; Hermundstad et al., 2014).
To construct a tractable model of the relation between natural odor statistics and olfactory receptor distributions, we describe the olfactory environment as a multivariate Gaussian with mean ${\mathbf{\mathbf{c}}}_{0}$ and covariance matrix $\mathrm{\Gamma}$,
This can be thought of as a maximumentropy approximation of the true distribution of odorant concentrations, constrained by the environmental means and covariances. This simple environmental model misses some sparse structure that is typical in olfactory scenes (Yu et al., 2015; Krishnamurthy et al., 2017). Nevertheless, approximating natural distributions with Gaussians is common in the efficientcoding literature, and often captures enough detail to be predictive (van Hateren, 1992a; van Hateren, 1992b; Van Hateren, 1993; Hermundstad et al., 2014). This may be because early sensory systems in animals are able to adapt more effectively to loworder statistics which are easily represented by neurons in their mean activity and pairwise correlations.
The number $N$ of odorants that we use to represent an environment need not be as large as the total number of possible volatile molecules. We can instead focus on only those odorants that are likely to be encountered at meaningful concentrations by the organism that we study, leading to a much smaller value for $N$. In practice, however, we are limited by the available receptor affinity data. Our quantitative analyses are generally based on data measured using panels of 110 odorants in fly (Hallem and Carlson, 2006) and 63 in mammals (Saito et al., 2009).
We next build a model for how the activity at the glomeruli depends on the olfactory environment. We work in an approximation in which the responses depend linearly on the concentration values:
where ${r}_{a}$ is the response of the glomerulus indexed by $a$, ${S}_{ai}$ is the expected response of a single sensory neuron expressing receptor type $a$ to a unit concentration of odorant $i$, and ${K}_{a}$ is the number of neurons of type $a$. The second term describes noise, with ${\eta}_{a}$, the noise for a single OSN, modeled as a Gaussian with mean 0 and standard deviation ${\sigma}_{a}$, ${\eta}_{a}\sim \mathcal{\mathcal{N}}(0,{\sigma}_{a}^{2})$.
The approximation we are using can be seen as linearizing the responses of olfactory sensory neurons around an operating point. This has been shown to accurately capture the response of olfactory receptors to odor mixtures in certain concentration ranges (Singh et al., 2018). While odor concentrations in natural scenes span many orders of magnitude and are unlikely to always stay within the linear regime, the effect of the nonlinearities on the information maximization procedure that we implement below is less strong (see Appendix 3 for a comparison between our linear approximation and a nonlinear, competitive binding model in a toy example). One advantage of employing the linear approximation is that it requires a minimal set of parameters (the sensing matrix coefficients $S}_{ai$), while nonlinear models in general require additional information (such as a Hill coefficient and a maximum activation for each receptorodorant pair for a competitive binding model; see Appendix 3).
Information maximization
We quantify the information that responses, $\mathbf{r}=({r}_{1},\dots ,{r}_{M})$, contain about the environment vector, $\mathbf{c}=({c}_{1},\dots ,{c}_{N})$, using the mutual information $I(\mathbf{r},\mathbf{c})$:
where $P(\mathbf{\mathbf{r}},\mathbf{\mathbf{c}})$ is the joint probability distribution over response and concentration vectors, $P(\mathbf{\mathbf{r}}\mathbf{\mathbf{c}})$ is the distribution of responses conditioned on the environment, and $P(\mathbf{\mathbf{r}})$ is the marginal distribution of the responses alone. Given our assumptions, all these distributions are Gaussian, and the integral can be evaluated analytically (see Appendix 2). The result is
where the overlap matrix $Q$ is related to the covariance matrix $\mathrm{\Gamma}$ of odorant concentrations (from Equation (1)),
and $\mathbb{K}$ and $\mathrm{\Sigma}$ are diagonal matrices of OSN abundances ${K}_{a}$ and noise variances ${\sigma}_{a}^{2}$, respectively:
The overlap matrix $Q$ is equal to the covariance matrix of OSN responses in the absence of noise (${\sigma}_{a}=0$; see Appendix 2). Thus, it is a measure of the strength of the usable olfactory signal. In contrast, the quantity $\mathrm{\Sigma}{\mathbb{K}}^{1}$ is a measure of the amount of noise in the responses, where the term $\mathbb{K}}^{1$ corresponds to the effect of averaging over OSNs of the same type. This implies that the quantity $\mathbb{K}{\mathrm{\Sigma}}^{1}Q$ is a measure of the signaltonoise ratio (SNR) in the system (more precisely, its square), so that Equation (4) represents a generalization to multiple, correlated channels of the classical result for a single Gaussian channel, $I=\frac{1}{2}\mathrm{log}\left(1+{\mathrm{S}\mathrm{N}\mathrm{R}}^{2}\right)$ (Shannon, 1948; van Hateren, 1992a; van Hateren, 1992b). In the linear approximation that we are using, the information transmitted through the system is the same whether all OSNs with the same receptor type converge to one or multiple glomeruli (see Appendix 2). Because of this, for convenience we take all neurons of a given type to converge onto a single glomerulus (Figure 1a).
The OSN numbers ${K}_{a}$ cannot grow without bound; they are constrained by the total number of neurons in the olfactory epithelium. Thus, to find the optimal distribution of receptor types, we maximize $I(\mathbf{\mathbf{r}},\mathbf{\mathbf{c}})$ with respect to $\{{K}_{a}\}$, subject to the constraints that: (1) the total number of receptor neurons is fixed (${\sum}_{a}{K}_{a}={K}_{\text{tot}}$); and (2) all neuron numbers are nonnegative:
Throughout the paper, we treat the OSN abundances ${K}_{a}$ as real numbers instead of integers, which is a good approximation as long as they are not very small. The optimization can be performed analytically using the KarushKuhnTucker (KKT) conditions (Boyd and Vandenberghe, 2004) (see Appendix 2), but in practice it is more convenient to use numerical optimization.
Note that in contrast to other work that has used information maximization to study the olfactory system (e.g. Zwicker et al., 2016), here we optimize over the OSN numbers ${K}_{a}$, while keeping the affinity profiles of the receptors (given by the sensing matrix elements ${S}_{ia}$) constant. Below we analyze how the optimal distribution of receptor types depends on receptor affinities, odor statistics, and the size of the olfactory epithelium.
Receptor diversity grows with OSN population size
Large OSN populations
In our model, receptor noise is reduced by averaging over the responses from many sensory neurons. As the number of neurons increases, ${K}_{\text{tot}}\to \mathrm{\infty}$, the signaltonoise ratio (SNR) becomes very large (see Equation (2)). When this happens, the optimization with respect to OSN numbers ${K}_{a}$ can be solved analytically (see Appendix 2), and we find that the optimal receptor distribution is given by
where $A$ is the inverse of the overlap matrix $Q$ from Equation (5), $A={Q}^{1}$, $\sigma}_{a}^{2$ are the receptor noise variances (Equation (6)), and $\overline{{\sigma}^{2}A}=\sum {\sigma}_{a}^{2}{A}_{aa}/M$ is a constant enforcing the constraint $\sum {K}_{a}={K}_{\mathrm{t}\mathrm{o}\mathrm{t}}$. When ${K}_{\text{tot}}$ is sufficiently large, the constant first term dominates, meaning that the receptor distribution is essentially uniform, with each receptor type being expressed in a roughly equal fraction of the total population of sensory neurons. In this limit, the receptor distribution is as even and as diverse as possible given the genetically encoded receptor types. The small differences in abundance are related to the diagonal elements of the inverse overlap matrix $A$, modulated by the noise variances ${\sigma}_{a}^{2}$ (Figure 2a). The information maximum in this regime is shallow because only a change in OSN numbers of order ${K}_{\text{tot}}/M$ can have a significant effect on the noise level for the activity of each glomerulus. Put another way, when the OSN numbers ${K}_{a}$ are very large, the glomerular responses are effectively noiseless, and the number of receptors of each type has little effect on the reliability of the responses. This scenario applies as long as the OSN abundances ${K}_{a}$ are much larger than the elements of the inverse overlap matrix $A$.
Small and intermediatesized OSN populations
When the number of neurons is very small, receptor noise can overwhelm the response to the environment. In this case, the best strategy is to focus all the available neurons on a single receptor type, thus reducing noise by summation as much as possible (Figure 2b). The receptor type that yields the most information will be the one whose response is most variable in natural scenes as compared to the amount of receptor noise; that is, the one that corresponds to the largest value of ${Q}_{aa}/{\sigma}_{a}^{2}$—see Appendix 2 for a derivation. This is reminiscent of a result in vision where the variance of a stimulus predicted its perceptual salience (Hermundstad et al., 2014).
As the total number of neurons increases, the added benefit of summing to lower noise for a single receptor type diminishes, and at some critical value it is more useful to populate a second receptor type that provides unique information not available in responses of the first type (Figure 2b). This process continues as the number of neurons increases, so that in an intermediate SNR range, where noise is significant but does not overwhelm the olfactory signal, our model leads to a highly nonuniform distribution of receptor types (see the trend in Figure 2b as the number of OSNs increases). Indeed, an inhomogeneous distribution of this kind is seen in mammals (IbarraSoria et al., 2017). Broadly, this is consistent with the idea that living systems conserve resources to the extent possible, and thus the number of OSNs (and therefore the SNR) will be selected to be in an intermediate range in which there are just enough to make all the available receptors useful.
Increasing OSN population size
Our model predicts that, all else being equal, the number of receptor types that are expressed should increase monotonically with the total number of sensory neurons, in a series of step transitions (see Figure 2c). Strictly speaking, this is a prediction that applies in a constant olfactory environment and with a fixed receptor repertoire; in terms of the parameters in our model, the total number of neurons ${K}_{\text{tot}}$ is varied while the sensing matrix $S$ and environmental statistics $\mathrm{\Gamma}$ stay the same. Keeping in mind that these conditions are not usually met by distinct species, we can nevertheless ask whether, broadly speaking, there is a relation between the number of functional receptor genes and the size of the olfactory epithelium in various species.
To this end, we looked at several mammals for which the number of OR genes and the size of the olfactory epithelium were measured (Figure 2f). We focused on the intact OR genes (Niimura et al., 2014), based on the expectation that receptor genes that tend to not be used are more likely to undergo deleterious mutations. We have not found many direct measurements of the number of neurons in the epithelium for different species, so we estimated this based on the area of the olfactory epithelium (Moulton, 1967; Pihlström et al., 2005; Gross et al., 1982; Smith et al., 2014). There is a known allometric scaling relation stating that the number of neurons per unit mass for a species decreases as the 0.3 power of the typical body mass (HerculanoHouzel et al., 2015). Assuming a fixed number of layers in the olfactory epithelial sheet, this implies that the number of neurons in the epithelium should scale as $N}_{\text{OSN}}\propto (\text{epithelial area})/(\text{body mass}{)}^{\frac{2}{3}\cdot 0.3$. We applied this relation to epithelial areas using the typical mass of several species (Rousseeuw and Leroy, 1987; FCI, 2018; Gross et al., 1982; Smith et al., 2014). The trend is consistent with expectations from our model (Figure 2f), keeping in mind uncertainties due to species differences in olfactory environments, receptor affinities, and behavior (e.g. consider marmoset vs. rat). A direct comparison is more complicated in insects, where even closely related species can vary widely in degree of specialization and thus can experience very different olfactory environments (Dekker et al., 2006). As we discuss below, our model’s detailed predictions can be more specifically tested in controlled experiments that measure the effect of a known change in odor environment on the olfactory receptor distributions of individual mammals, as in IbarraSoria et al. (2017).
Optimal OSN abundances are contextdependent
We can predict the optimal distribution of receptor types given the sensing matrix $S$ and the statistics of odors by maximizing the mutual information in Equation (4) while keeping the total number of neurons ${K}_{\text{tot}}={\sum}_{a}{K}_{a}$ constant. We tested the effect of changing the variance of a single odorant, and found that the effect on the optimal receptor abundances depends on the context of the background olfactory environment. Increased exposure to a particular ligand can lead to increased abundance of a given receptor type in one context, but to decreased abundance in another (Figure 3). In fact, patterns of this kind have been reported in recent experiments (Santoro and Dulac, 2012; Zhao et al., 2013; Cadiou et al., 2014; IbarraSoria et al., 2017). To understand this contextdependence better, we analyzed the predictions of our model in various signal and noise scenarios.
One factor that does not affect the optimal receptor distribution in our model is the average concentration vector ${\mathbf{\mathbf{c}}}_{0}$. This is because it corresponds to odors that are always present and therefore offer no new information about the environment. This is consistent with experiment (IbarraSoria et al., 2017), where it was observed that chronic odor exposure does not affect receptor abundances in the epithelium. In the rest of the paper, we thus restrict our attention to the covariance matrix of odorant concentrations, $\mathrm{\Gamma}$.
The problem of maximizing the amount of information that OSN responses convey about the odor environment simplifies considerably if these responses are weakly correlated. In this case, standard efficient coding theory says that receptors whose activities fluctuate more extensively in response to the olfactory environment provide more information to brain, while receptors that are active at a constant rate or are very noisy provide less information. In this circumstance, neurons expressing receptors with large signaltonoise ratio (SNR, i.e. signal variance as compared to noise variance) should increase in proportion relative to neurons with low signaltonoise ratio (see Appendix 2 for a derivation). In terms of our model, the signal variance of glomerular responses is given by diagonal elements of the overlap matrix $Q$ (Equation 5), while the noise variance is ${\sigma}_{a}^{2}$; so we expect ${K}_{a}$, the number of OSNs of type $a$, to increase with ${Q}_{aa}/{\sigma}_{a}^{2}$. Responses are less correlated if receptors are narrowly tuned, and we find indeed that if each receptor type responds to only a small number of odorants, the abundances of OSNs of each type correlate well with their variability in the environment (narrowtuning side of Figure 2d). This is also consistent with the results at high SNR: we saw above that in that case ${K}_{a}\approx C{\sigma}_{a}^{2}{({Q}^{1})}_{aa}$, and when response correlations are weak, $Q$ is approximately diagonal, and thus ${({Q}^{1})}_{aa}\approx 1/{Q}_{aa}$.
The biological setting is better described in terms of widely tuned sensing matrices (Hallem and Carlson, 2006), and an intermediate SNR level in which noise is important, but does not dominate the responses of most receptors. We therefore generated sensing matrices with varying tuning width by changing the number of odorants that elicit strong activity in each receptor (as detailed in Appendix 1). We found that as receptors begin responding to a greater diversity of odorants, the correlation structure of their activity becomes important in determining the optimal receptor distribution; it is no longer sufficient to just examine the signal to noise ratios of each receptor type separately as a conventional theory suggests (widetuning side of Figure 2d). In other words, the optimal abundance of a receptor type depends not just on its activity level, but also on the context of the correlated activity levels of all the other receptor types. These correlations are determined by the covariance structures of the environment and of the sensing matrix.
In fact, across the range of tuning widths the optimal receptor abundances ${K}_{a}$ are correlated with the inverse of the overlap matrix, $A={Q}^{1}$ (Figure 2e). For narrow tuning widths, the overlap matrix $Q$ is approximately diagonal (because correlations between receptors are weak) and so ${Q}^{1}$ is simply the matrix of the inverse diagonal elements of $Q$. Thus, in this limit, the correlation with ${Q}^{1}$ simply follows from the correlation with $Q$ that we discussed above. As the tuning width increases keeping the total number of OSNs ${K}_{\text{tot}}$ constant, the responses from each receptor grow stronger, increasing the SNR, even as the offdiagonal elements of the overlap matrix $Q$ become significant. In the limit of high SNR, the analytical formula ${K}_{a}\approx C{\sigma}_{a}^{2}{Q}_{aa}^{1}$ (Equation 8) ensures that the OSN numbers ${K}_{a}$ are still correlated with the diagonal elements of ${Q}^{1}$, despite the presence of large offdiagonal components. Because of the matrix inversion in ${Q}^{1}$, the optimal abundance for each receptor type is affected in this case by the full covariance structure of all the responses and not just by the variance ${Q}_{aa}$ of the receptor itself. Mathematically, this is because the diagonal elements of ${Q}^{1}$ are functions of all the variances and covariances in the overlap matrix $Q$. This dependence of each abundance on the full covariance translates to a complex contextdependence whereby changing the same ligand in different background environments can lead to very different adapted distributions of receptors. In Appendix 6 we show that the correlation with the inverse overlap matrix has an intuitive interpretation: receptors which either do not fluctuate much or whose values can be guessed based on the responses of other receptors should have low abundances.
Environmental changes lead to complex patterns of OSN abundance changes
To investigate how the structure of the optimal receptor repertoire varies with the olfactory environment, we first constructed a background in which the concentrations of 110 odorants were distributed according to a Gaussian with a randomly chosen covariance matrix (e.g., Figure 4a; see Appendix 4 for details). From this base, we generated two different environments by adding a large variance to 10 odorants in environment 1, and to 10 different odorants in environment 2 (Figure 4b). We then considered the optimal distribution in these environments for a repertoire of 24 receptor types with odor affinities inferred from (Hallem and Carlson, 2006). We found that when the number of olfactory sensory neurons ${K}_{\text{tot}}$ is large, and thus the signaltonoise ratio is high, the change in odor statistics has little effect on the distribution of receptors (Figure 4c). This is because at high SNR, all the receptors are expressed nearly uniformly as discussed above, and this is true in any environment. When the number of neurons is smaller (or, equivalently, the signaltonoise ratio is in a low or intermediate regime), the change in environment has a significant effect on the receptor distribution, with some receptor types becoming more abundant, others becoming less abundant, and yet others not changing much between the environments (see Figure 4d). This mimics the kinds of complex effects seen in experiments in mammals (Schwob et al., 1992; Santoro and Dulac, 2012; Zhao et al., 2013; Dias and Ressler, 2014; Cadiou et al., 2014; IbarraSoria et al., 2017).
Changing odor identities has more extreme effects on receptor distributions than changing concentrations
In the comparison above, the two environment covariance matrices differed by a large amount for a small number of odors. We next compared environments with two different randomly generated covariance matrices, each generated in the same way as the background environment in Figure 4a. The resulting covariance matrices (Figure 5a, top) are very different in detail (the correlation coefficient between their entries is close to zero; distribution of changes in Figure 5b, red line), although they look similar by eye. Despite the large change in the detailed structure of the olfactory environment, the corresponding change in optimal receptor distribution is typically small, with a small fraction of receptor types experiencing large changes in abundance (red curve in Figure 5c). The average abundance of each receptor in these simulations was about 1000, and about 90% of all the abundance change values $\left\mathrm{\Delta}{K}_{i}\right$ were below 20% of this, which is the range shown on the plot in Figure 5c. Larger changes also occurred, but very rarely: about 0.1% of the abundance changes were over 800.
If we instead engineer two environments that are almost nonoverlapping so that each odorant is either common in environment 1, or in environment 2, but not in both (Figure 5a, bottom; see Appendix 4 for how this was done), the changes in optimal receptor abundances between environments shift away from midrange values towards higher values (blue curve in Figure 5c). For instance, 40% of abundance changes lie in the range $\left\mathrm{\Delta}K\right\phantom{\rule{thinmathspace}{0ex}}>\phantom{\rule{thinmathspace}{0ex}}50$ in the nonoverlapping case, while the proportion is 28% in the generic case.
It seems intuitive that animals that experience very different kinds of odors should have more striking differences in their receptor repertoires than those that merely experience the same odors with different frequencies. Intriguingly, however, our simulations suggest that the situation may be reversed at the very low end: the fraction of receptors for which the predicted abundance change is below 0.1, $\left\mathrm{\Delta}K\right\phantom{\rule{thinmathspace}{0ex}}<\phantom{\rule{thinmathspace}{0ex}}0.1$, is about 2% in the generic case but over 9% for nonoverlapping environment pairs. Thus, changing between nonoverlapping environments emphasizes the more extreme changes in receptor abundances, either the ones that are close to zero or the ones that are large. In contrast, a generic change in the environment leads to a more uniform distribution of abundance changes. Put differently, the particular way in which the environment changes, and not only the magnitude of the change, can affect the receptor distribution in unexpected ways.
The magnitude of the effect of environmental changes on the optimal olfactory receptor distribution is partly controlled by the tuning of the olfactory receptors (Figure 5d). If receptors are narrowly tuned, with each type responding to a small number of odorants, changes in the environment tend to have more drastic effects on the receptor distribution than when the receptors are broadly tuned (Figure 5d), an effect that could be experimentally tested.
Model predictions qualitatively match experiments
Our study opens the exciting possibility of a causal test of the hypothesis of efficient coding in sensory systems, where a perturbation in the odor environment could lead to predictable adaptations of the olfactory receptor distribution during the lifetime of an individual. This does not happen in insects, but it can happen in mammals, since their receptor neurons regularly undergo apoptosis and are replaced.
A recent study demonstrated reproducible changes in olfactory receptor distributions of the sort that we predict in mice (IbarraSoria et al., 2017). These authors raised two groups of mice in similar conditions, exposing one group to a mixture of four odorants (acetophenone, eugenol, heptanal, and Rcarvone) either continuously or intermittently (by adding the mixture to their water supply). Continuous exposure to the odorants had no effect on the receptor distribution, in agreement with the predictions of our model. In contrast, intermittent exposure did lead to systematic changes (Figure 6a).
We used our model to run an experiment similar to that of IbarraSoria et al. (2017) in silico (Figure 6b). Using a sensing matrix based on odor response curves for mouse and human receptors (data for 59 receptors from Saito et al. (2009)), we calculated the predicted change in OSN abundances between two different environments with random covariance matrices constructed as described above. We ran the simulations 24 times, modifying the odor environments each time by adding a small amount of Gaussian random noise to the square roots of these covariance matrices to model small perturbations (details in Appendix 4; range bars in Figure 6b). The results show that the abundances of already numerous receptors do not change much, while there is more change for less numerous receptors. The frequencies of rare receptors can change dramatically, but are also more sensitive to perturbations of the environment (large range bars in Figure 6b).
These results qualitatively match experiment (Figure 6a), where we see the same pattern of the largest reproducible changes occurring for receptors with intermediate abundances. The experimental data is based on receptor abundance measured by RNAseq which is a proxy for counting OSN numbers (IbarraSoria et al., 2017). In our model, the distinction between receptor numbers and OSN numbers is immaterial because a change in the number of receptors expressed per neuron has the same effect as a change in neuron numbers. In general, additional experiments are needed to measure both the number of receptors per neuron and the number of neurons for each receptor type.
A framework for a quantitative test
Given detailed information regarding the affinities of olfactory receptors, the statistics of the odor environment, and the size of the olfactory epithelium (through the total number of neurons ${K}_{\text{tot}}$), our model makes fully quantitative predictions for the abundances of each OSN type. Existing experiments (e.g. IbarraSoria et al., 2017) do not record necessary details regarding the odor environment of the control group and the magnitude of the perturbation experienced by the exposed group. However, such data can be collected using available experimental techniques. Anticipating future experiments, we provide a Matlab (RRID:SCR_001622) script on GitHub (RRID:SCR_002630) to calculate predicted OSN numbers from our model given experimentallymeasured sensing parameters and environment covariance matrix elements (https://github.com/ttesileanu/OlfactoryReceptorDistribution).
Given the huge number of possible odorants (Yu et al., 2015), the sensing matrix of affinities between all receptor types in a species and all environmentally relevant odorants is difficult to measure. One might worry that this poses a challenge for our modeling framework. One approach might be to use lowdimensional representations of olfactory space (e.g. Koulakov et al., 2011; Snitz et al., 2013), but there is not yet a consensus on the sufficiency of such representations. For now, we can ask how the predictions of our model change upon subsampling: if we only know the responses of a subset of receptors to a subset of odorants, can we still accurately predict the OSN numbers for the receptor types that we do have data for? Figure 7a and b show that such partial data do lead to robust statistical predictions of overall receptor abundances.
First steps toward a dynamical model in mammals
We have explored the structure of olfactory receptor distributions that code odors efficiently, that is are adapted to maximize the amount of information that the brain gets about odors. The full solution to the optimization problem, Equation (7), depends in a complicated nonlinear way on the receptor affinities $S$ and covariance of odorant concentrations $\mathrm{\Gamma}$. The distribution of olfactory receptors in the mammalian epithelium, however, must arise dynamically from the pattern of apoptosis and neurogenesis (Calof et al., 1996). At a qualitative level, in the efficient coding paradigm that we propose, the receptor distribution is related to the statistics of natural odors, so that the life cycle of neurons would have to depend dynamically on olfactory experience. Such modulation of OSN lifetime by exposure to odors has been observed experimentally (Santoro and Dulac, 2012; Zhao et al., 2013) and could, for example, be mediated by feedback from the bulb (Schwob et al., 1992).
To obtain a dynamical model, we started with a gradient ascent algorithm for changing receptor numbers, and modified it slightly to impose the constraints that OSN numbers are nonnegative, ${K}_{a}\ge 0$, and their sum ${K}_{\text{tot}}={\sum}_{a}{K}_{a}$ is bounded (details in Appendix 5). This gives
where $\alpha $ is a learning rate, ${\sigma}_{a}^{2}$ is the noise variance for receptor type $a$, and $R$ is the covariance matrix of glomerular responses,
with the angle brackets denoting ensemble averaging over both odors and receptor noise. In the absence of the experiencerelated term ${({R}^{1})}_{aa}$, the dynamics from Equation (9) would be simply logistic growth: the population of OSNs of type $a$ would initially grow at a rate $\alpha $, but would saturate when ${K}_{a}=1/\lambda $ because of the populationdependent death rate $\lambda {K}_{a}$. In other words, the quantity $M/\lambda $ sets the asymptotic value for the total population of sensory neurons, ${K}_{\text{tot}}\to M/\lambda $, with $M$ being the number of receptor types.
Because of the last term in Equation (9), the death rate in our model is influenced by olfactory experience in a receptordependent way. In contrast, the birth rate is not experiencedependent and is the same for all OSN types. Indeed, in experiments, the odor environment is seen to have little effect on receptor choice, but does modulate the rate of apoptosis in the olfactory epithelium (Santoro and Dulac, 2012). Our results suggest that, if olfactory sensory neuron lifetimes are appropriately anticorrelated with the inverse response covariance matrix, then the receptor distribution in the epithelium can converge to achieve optimal information transfer to the brain.
The elements of the response covariance matrix ${R}_{ab}$ could be estimated by temporal averaging of cooccurring glomerular activations via lateral connections between glomeruli (Mori et al., 1999). Performing the inverse necessary for our model is more intricate. The computations could perhaps be done by circuits in the bulb and then fed back to the epithelium through known mechanisms (Schwob et al., 1992),
Within our model, Figure 8a shows an example of receptor numbers converging to the optimum from random initial values. The sensing matrix used here is based on mammalian data (Saito et al., 2009) and we set the total OSN number to ${K}_{\text{tot}}=2000$. The environment covariance matrix is generated using the random procedure described earlier (details in Appendix 4). We see that some receptor types take longer than others to converge (the time axis is logarithmic, which helps visualize the whole range of convergence behaviors). Roughly speaking, convergence is slower when the final OSN abundance is small, which is related to the fact that the rate of change $d{K}_{a}/dt$ in Equation (9) vanishes in the limit ${K}_{a}\to 0$. For the same reason, OSN populations that start at a very low level also take a long time to converge.
In Figure 8b, we show convergence to the same final state, but this time starting from a distribution that is not random but was optimized for a different environment. The initial and final environments are the same as the two environments used in the previous section to compare the simulations to experimental findings (Figure 6b). Interestingly, many receptor types actually take longer to converge in this case compared to the random starting point, perhaps because there are local optima in the landscape of receptor distributions. Given such local minima, stochastic fluctuations will allow the dynamics to reach the global optimum more easily. In realistic situations, there are many sources of such variability, for example, sampling noise due to the fact that the response covariance matrix $R$ must be estimated through stochastic odor encounters and noisy receptor readings. In fact, in Figure 8b, we added a small amount of noise (corresponding to $\pm 0.05{K}_{\text{tot}}/M$) to the initial distribution of receptors to improve convergence rates.
Discussion
We built a model for the distribution of receptor types in the olfactory epithelium that is based on efficient coding, and assumes that the abundances of different receptor types are adapted to the statistics of natural odors in a way that maximizes the amount of information conveyed to the brain by glomerular responses. This model predicts a nonuniform distribution of receptor types in the olfactory epithelium, as well as reproducible changes in the receptor distribution after perturbations to the odor environment. In contrast to other applications of efficient coding, our model operates in a regime in which there are significant correlations between sensors because the adaptation of OSN abundances occurs upstream of the brain circuitry that can decorrelate olfactory responses. In this regime, OSN abundances depend on the full correlation structure of the inputs, leading to predictions that are contextdependent in the sense that whether the abundance of a specific receptor type goes up or down due to a shift in the environment depends on the global context of the responses of all the other receptors. All these striking phenomena have been observed in recent experiments and had not been explained prior to this study.
In our framework, the sensitivity of the receptor distribution to changes in odor statistics is affected by the tuning of the olfactory receptors, with narrowly tuned receptors being more readily affected by such changes than broadly tuned ones. The model also predicts that environments that differ in the identity of the odors that are present will lead to greater deviations in the optimal receptor distribution than environments that differ only in the statistics with which these odors are encountered. Likewise, the model broadly predicts a monotonic relationship between the number of receptor types found in the epithelium and the total number of olfactory sensory neurons, all else being equal.
A detailed test of our model requires more comprehensive measurements of olfactory environments than are currently available. Our hope is that studies such as ours will spur interest in measuring the natural statistics of odors, opening the door for a variety of theoretical advances in olfaction, similar to what was done for vision and audition. Such measurements could for instance be performed by using mass spectrometry to measure the chemical composition of typical odor scenes. Given such data, and a library of receptor affinities, our GitHub (RRID:SCR_002630) online repository provides an easytouse script that uses our model to predict OSN abundances. For mammals, controlled changes in environments similar to those in IbarraSoria et al. (2017) could provide an even more stringent test for our framework.
To our knowledge, this is the first time that efficient coding ideas have been used to explain the pattern of usage of receptors in the olfactory epithelium. Our work can be extended in several ways. OSN responses can manifest complex, nonlinear responses to odor mixtures. Accurate models for how neurons in the olfactory epithelium respond to complex mixtures of odorants are just starting to be developed (e.g. Singh et al., 2018), and these can in principle be incorporated in an informationmaximization procedure similar to ours. More realistic descriptions of natural odor environments can also be added, as they amount to changing the environmental distribution $P(\mathbf{\mathbf{c}})$. For example, the distribution of odorants could be modeled using a Gaussian mixture, rather than the normal distribution used in this paper to enable analytic calculations. Each Gaussian in the mixture would model a different odor object in the environment, more closely approximating the sparse nature of olfactory scenes discussed in, for example, Krishnamurthy et al. (2017).
Of course, the goal of the olfactory system is not simply to encode odors in a way that is optimal for decoding the concentrations of volatile molecules in the environment, but rather to provide an encoding that is most useful for guiding future behavior. This means that the value of different odors might be an important component shaping the neural circuits of the olfactory system. In applications of efficient coding to vision and audition, maximizing mutual information, as we did, has proved effective even in the absence of a treatment of value (Laughlin, 1981; Atick and Redlich, 1990; van Hateren, 1992a; Olshausen and Field, 1996; Simoncelli and Olshausen, 2001; Fairhall et al., 2001; Lewicki, 2002; Ratliff et al., 2010; Garrigan et al., 2010; Tkacik et al., 2010; Hermundstad et al., 2014; Palmer et al., 2015; Salisbury and Palmer, 2016). However, in general, understanding the role of value in shaping neural circuits is an important experimental and theoretical problem. To extend our model in this direction, we would replace the mutual information between odorant concentrations and glomerular responses by a different function that takes into account value assignments (see, e.g. Rivoire and Leibler, 2011). It could be argued, though, that such specialization to the most behaviorally relevant stimuli might be unnecessary or even counterproductive close to the sensory periphery. Indeed, a highly specialized olfactory system might be better at reacting to known stimuli, but would be vulnerable to adversarial attacks in which other organisms take advantage of blind spots in coverage. Because of this, and because precise information regarding how different animals assign value to different odors is scarce, we leave these considerations for future work.
One exciting possibility suggested by our model is a way to perform a first causal test of the efficient coding hypothesis for sensory coding. Given sufficiently detailed information regarding receptor affinities and natural odor statistics, experiments could be designed that perturb the environment in specified ways, and then measure the change in olfactory receptor distributions. Comparing the results to the changes predicted by our theory would provide a strong test of efficient coding by early sensory systems in the brain.
Materials and methods
Software and data
Request a detailed protocolThe code (written in Matlab, RRID:SCR_001622) and data that we used to generate all the results and figures in the paper is available on GitHub (RRID:SCR_002630), at https://github.com/ttesileanu/OlfactoryReceptorDistribution (Teşileanu, 2019; copy archived at https://github.com/elifesciencespublications/OlfactoryReceptorDistribution).
Appendix 1
Choice of sensing matrices and receptor noise variances
We used three types of sensing matrices in this study. Two were based on experimental data, one using fly receptors (Hallem and Carlson, 2006), and one using mouse and human receptors (Saito et al., 2009); and another type of sensing matrix was based on randomlygenerated receptor affinity profiles. These can all be either directly downloaded from our repository on GitHub (RRID:SCR_002630), https://github.com/ttesileanu/OlfactoryReceptorDistribution, or generated using the code available there.
Fly sensing matrix
Some of our simulations used a sensing matrix based on Drosophila receptor affinities, as measured by Hallem and Carlson (Hallem and Carlson, 2006). This includes the responses of 24 of the 60 receptor types in the fly against a panel of 110 odorants, measured using singleunit electrophysiology in a mutant antennal neuron. We used the values from Table S1 in (Hallem and Carlson, 2006) for the sensing matrix elements. To estimate receptor noise, we used the standard deviation measured for the background firing rates for each receptor (data obtained from the authors). The fly data has the advantage of being more complete than equivalent datasets in mammals.
Mammalian sensing matrix
When comparing our model to experimental findings from (IbarraSoria et al., 2017), we used a sensing matrix based on mouse and human receptor affinity data from (Saito et al., 2009). This was measured using heterologous expression of olfactory genes, and tested in total 219 mouse and 245 human receptor types against 93 different odorants. However, only 49 mouse receptors and 10 human receptors exhibited detectable responses against any of the odorants, while only 63 odorants activated any receptors. From the remaining 59 × 63 = 3717 receptor–odorant pairs, only 335 (about 9%) showed a response, and were assayed at 11 different concentration points. In this paper, we used the values obtained for the highest concentration (3 mM).
Random sensing matrices
The random sensing matrices matrices used in the main text (and referred to as ‘tuning’ in some of the figures in this Appendix) were generated as follows. We started by treating the column (i.e. odorant) index as a onedimensional odor coordinate with periodic boundary conditions. We normalized the index to a coordinate $x$ running from 0 to 1. For each receptor, we then chose a center ${x}_{0}$ along this line, corresponding to the odorant to which the receptor has maximum affinity, and a standard deviation $\sigma $, corresponding to the tuning width of the receptor. Note that both ${x}_{0}$ and $\sigma $ are allowed to be real numbers, so that the maximum affinity can occur at a position that does not correspond to any particular odorant from the sensing matrix.
To obtain a belllike response profile for the receptors while preserving the periodicity of the odor coordinate we chose, we defined the response affinity to odorant $x$ by
This expression can be obtained by imagining odorant space as a circle embedded in a twodimensional plane, with odorant $x$ mapped to an angle $\theta =2\pi x$ on this circle, and considering a Gaussian response profile in this twodimensional embedding space. This is simply a convenient choice for treating odor space in a way that eliminates artifacts at the edges of the sensing matrix, and we do not assign any significance to the particular coordinate system that we used.
The centers ${x}_{0}$ for the Gaussian profiles for each of the receptors were chosen uniformly at random, and the tuning width $\sigma $ was either a fixed parameter for the entire sensing matrix, or was uniformly sampled from an interval. Before using the matrices we randomly shuffled the columns to remove the dependencies between neighboring odorants, and finally added some amount of random Gaussian noise (mean centered and with standard deviation 1/200). The overall scale of the sensing matrices was set by multiplying all the affinities by 100, which yielded values comparable to the measured firing rates in fly olfactory neurons (Hallem and Carlson, 2006).
For the robustness results below we also generated random matrices in additional ways: (1) ‘gaussian’: drawing the affinities from a Gaussian distribution (with zero mean and standard deviation 2), (2) ‘bernoulli’: drawing from a Bernoulli distribution (with elements equal to 5 with probability 30%, and 0 with probability 70%), (3) ‘signed’: drawing from a Bernoulli distribution followed by choosing the sign (so that elements are 5 with probability 15%, –5 with probability 15%, and 0 with probability 70%); and (4, 5) ‘fly scrambled’ and ‘mammal scrambled’: scrambling the elements in the fly and mammalian datasets (across both odorants and receptors).
Robustness of results to changing the sensing matrix
Our qualitative results are robust across a variety of different choices for the sensing matrix (Appendix 1—figure 1). For instance, the optimal number of receptor types expressed in a fraction of the OSN population larger than 1% grows monotonically with the total number of neurons (Appendix 1—figure 2). Similarly, the general effect that environment change has on optimal OSN numbers, with less abundant receptor types changing more than more abundant ones, is generic across different choices of sensing matrices (Appendix 1—figure 3).
Appendix 2
Mathematical derivations
Deriving the expression for the mutual information
In the main text we assume a Gaussian distribution for odorant concentrations and approximate receptor responses as linear with additive Gaussian noise, Equation (2). Thus it follows that the marginal distribution of receptor responses is also Gaussian. Taking averages of the responses, $\u27e8{r}_{a}\u27e9$, and of products of responses, $\u27e8{r}_{a}{r}_{b}\u27e9$, over both the noise distribution and the odorant distribution, and using Equation (2) from the main text, we get a normal distribution of responses:
where the mean response vector ${\mathbf{\mathbf{r}}}_{0}$ and the response covariance matrix $R$ are given by
where $S$ is the sensing matrix, $\mathbb{K}$ is a diagonal matrix of OSN abundances, and $\mathrm{\Sigma}$ is the covariance matrix of receptor noises, $\mathrm{\Sigma}=\mathrm{d}\mathrm{i}\mathrm{a}\mathrm{g}({\sigma}_{1}^{2},\dots ,{\sigma}_{M}^{2})$ (also see the main text). Here, as in Equation (1) in the main text, ${\mathbf{\mathbf{c}}}_{0}$ is the mean concentration vector, $\mathrm{\Gamma}$ is the covariance matrix of odorant concentrations, and we use the overlap matrix from Equation (5) in the main text, $Q=S\mathrm{\Gamma}{S}^{T}$. Note that in the absence of noise ($\mathrm{\Sigma}=0$), the response matrix is simply the overlap matrix $Q$ modulated by the number of OSNs of each type, $R}_{\text{noiseless}}=\mathbb{K}Q\mathbb{K$.
The joint probability distribution over responses and concentrations, $P(\mathbf{\mathbf{r}},\mathbf{\mathbf{c}})$, is itself Gaussian. To calculate the corresponding covariance matrix, we need the covariances between responses, $\u27e8{r}_{a}{r}_{b}\u27e9\u27e8{r}_{a}\u27e9\u27e8{r}_{b}\u27e9$, which are just the elements of the response matrix $R$ from Equation (13) above; and between concentrations, $\u27e8{c}_{i}{c}_{j}\u27e9\u27e8{c}_{i}\u27e9\u27e8{c}_{j}\u27e9$, which are the elements of the environment covariance matrix $\mathrm{\Gamma}$, Equation (1) in the main text. In addition, we need the covariances between responses and concentrations, $\u27e8{r}_{a}{c}_{i}\u27e9\u27e8{r}_{a}\u27e9\u27e8{c}_{i}\u27e9$, which can be calculated using Equation (2) from the main text. We get:
with
The mutual information between responses and odors is then given by (see below for a derivation):
From Equation (13) we have
and from Equation (15),
where we used Equation (13) again, and employed Schur’s determinant identity (derived below). Thus,
This recovers the result quoted in the main text, Equation (4).
By using the fact that the diagonal matrices $\mathbb{K}$ and ${\mathrm{\Sigma}}^{1}$ commute, we can also write:
This shows that the mutual information can be written in terms of a symmetric ‘SNR matrix’ $\stackrel{~}{Q}={\mathrm{\Sigma}}^{1/2}Q{\mathrm{\Sigma}}^{1/2}$. This is simply the covariance matrix of responses in which each response was normalized by the noise variance of the corresponding receptor.
Schur’s determinant identity
The identity for the determinant of a 2 × 2 block matrix that we used in Equation (18) above can be derived in the following way. First, note that
Now, from the definition of the determinant it can be seen that
since all the products involving elements from the offdiagonal blocks must necessarily also involve elements from the 0 matrix. Thus, taking the determinant of Equation (21), we get the desired identity
Mutual information for Gaussian distributions
The expression from Equation (16) for the mutual information $I(\mathbf{\mathbf{r}},\mathbf{\mathbf{c}})$ can be derived by starting with the fact that $I$ is equal to the KullbackLeibler (KL) divergence from the joint distribution $P(\mathbf{\mathbf{r}},\mathbf{\mathbf{c}})$ to the product distribution $P(\mathbf{\mathbf{r}})P(\mathbf{\mathbf{c}})$. As a first step, let us calculate the KL divergence between two multivariate normals in $n$ dimensions:
where
Plugging the distribution functions into the logarithm, we have
where the normalization property of $p(\mathbf{\mathbf{x}})$ was used. Using also the definition of the mean and of the covariance matrix, we have
which implies
for any vector $\mu $ and matrix $C$. Plugging this into Equation (26), we get
We can now return to calculating the KL divergence from $P(\mathbf{\mathbf{r}},\mathbf{\mathbf{c}})$ to $P(\mathbf{\mathbf{r}})P(\mathbf{\mathbf{c}})$. Note that, since $P(\mathbf{\mathbf{r}})$ and $P(\mathbf{\mathbf{c}})$ are just the marginals of the joint distribution, the means of the variables are the same in the joint and in the product, so that the last term in the KL divergence vanishes. The covariance matrix for the product distribution is
so the product inside the trace becomes
where the entries replaced by '$\mathrm{\dots}$' need not be calculated because they drop out when the trace is taken. The sum of the dimensions of $R$ and $\mathrm{\Gamma}$ is equal to the dimension, $n$, of $\mathrm{\Lambda}$, so that the term involving the trace from Equation (29) also drops out, leaving us with the final result:
which is the same as Equation (16) that was used in the previous section.
Deriving the KKT conditions for the information optimum
In order to find the optimal distribution of olfactory receptors, we must maximize the mutual information from Equation (4) in the main text, subject to constraints. Let us first calculate the gradient of the mutual information with respect to the receptor numbers:
The cyclic property of the trace allows us to use the usual rules to differentiate under the trace operator, so we get
We now have to address the constraints. We have two kinds of constraints: an equality constraint that sets the total number of neurons, $\sum {K}_{a}={K}_{\text{tot}}$; and inequality constraints that ensure that all receptor abundances are nonnegative, ${K}_{a}\ge 0$. This can be done using the KarushKuhnTucker (KKT) conditions, which require the introduction of Lagrange multipliers: $\lambda $ for the equality constraint, and ${\mu}_{a}$ for the inequality constraints. At the optimum, we must have:
where the Lagrange multipliers for the inequality constraints, ${\mu}_{a}$, must be nonnegative, and must vanish unless the inequality is saturated:
Put differently, if ${K}_{a}\phantom{\rule{thinmathspace}{0ex}}>\phantom{\rule{thinmathspace}{0ex}}0$, then ${\mu}_{a}=0$ and $\partial I/\partial {K}_{a}=\lambda /2$; while if ${K}_{a}=0$, then $\partial I/\partial {K}_{a}=\lambda /2{\mu}_{a}\le \lambda /2$. Combined with Equation (34), this yields
The magnitude of $\lambda $ is set by imposing the normalization condition $\sum {K}_{a}={K}_{\text{tot}}$.
The manyneuron approximation
Suppose we are in the regime in which the total number of neurons is large, and in particular, each of the abundances ${K}_{a}$ is large. Then we can perform an expansion of the expression appearing in the KKT equations from Equation (37):
whose $aa$ component is
where we used $\stackrel{~}{Q}={\mathrm{\Sigma}}^{1/2}Q{\mathrm{\Sigma}}^{1/2}$. With the notation
we can plug into Equation (37) and get
This quadratic equation has only one large solution, and it is given approximately by
Combined with the normalization constraint, ${\sum}_{a}{K}_{a}={K}_{\text{tot}}$, this recovers Equation (8) from the main text.
Optimal distribution for uncorrelated responses
When the overlap matrix $Q=S\mathrm{\Gamma}{S}^{T}$ is diagonal, the optimization problem simplifies considerably. By plugging $Q=\mathrm{diag}({Q}_{aa})$ into Equation (4) in the main text, we find
We can again use the KKT approach and add Lagrange multipliers $\lambda $ and ${\mu}_{a}$ for enforcing the equality and inequality constraints, respectively,
and take derivatives with respect to ${K}_{a}$ to find the optimum,
with the condition that ${\mu}_{a}\ge 0$ and either ${\mu}_{a}$ or ${K}_{a}$ must vanish, ${\mu}_{a}{K}_{a}=0$. This leads to
showing that receptor abundances grow monotonically with ${Q}_{aa}/{\sigma}_{a}^{2}$. This explains the correlation between OSN abundances ${K}_{a}$ and receptor SNRs ${Q}_{aa}/{\sigma}_{a}^{2}$ when the responses are uncorrelated or weakly correlated.
First receptor type to be activated
When there is only one active receptor, ${K}_{x}={K}_{\text{tot}}$, ${K}_{a\ne x}=0$, the KKT conditions from Equation (37) are automatically satisfied. The receptor that is activated first can be found in this case by calculating the information $I(\mathbf{\mathbf{r}},\mathbf{\mathbf{c}})$ using Equation (4) from the main text while assuming an arbitrary index $x$ for the active receptor, and then finding $x={x}^{*}$ that yields the maximum value. Without loss of generality, we can permute the receptor indices such that $x=1$. Using Equation (19) and setting ${K}_{1}={K}_{\text{tot}}$, we have:
Thus, in general, the information when only receptor type $x$ is activated is given by
which implies that information is maximized when $x$ matches the receptor corresponding to the highest ratio between the diagonal value of the overlap matrix $Q$ and the receptor variance in that channel ${\sigma}_{x}^{2}$; that is the receptor that maximizes the signaltonoise ratio:
Another way to think of this result is by employing the usual expression for the capacity of a single Gaussian channel, and then finding the channel that maximizes this capacity.
Invariance of mutual information under invertible and differentiable transformations
Consider the mutual information between two variables $\mathbf{r}\in {\mathbb{R}}^{M}$ and $\mathbf{c}\in {\mathbb{R}}^{N}$:
Let us now define two different variables that depend on $\mathbf{\mathbf{r}}$ and $\mathbf{\mathbf{c}}$ in an invertible and continuouslydifferentiable (but in general nonlinear) way,
The joint probability distribution for the new variables is related to the joint distribution of the original variables through the Jacobian determinants,
where
For the marginals, we have
where we used the standard substitution formula for multiple integrals. We can now calculate the mutual information between the new variables:
Thus, invertible and continuouslydifferentiable transformations of either the response variables $\mathbf{\mathbf{r}}$ or the concentration variables $\mathbf{\mathbf{c}}$ in our model leave the mutual information unchanged.
Multiple glomeruli with the same affinity profile
In mammals, the axons from neurons expressing a given receptor type can project to anywhere from 2 to 16 different glomeruli. Here we show that in our setup, information transfer only depends on the total number of neurons of a given type, and not on the number of glomeruli to which they project.
The key observation is that mutual information, Equation (3) in the main text, is unchanged when the responses and/or concentrations are modified by invertible transformations (see previous section). In particular, linear transformations of the responses do not affect the information values. Suppose that we have a case in which two receptors $p$ and $q$ have identical affinities, so that ${S}_{pi}={S}_{qi}$ for all odorants $i$. We can then form linear combinations of the corresponding glomerular responses,
and consider a transformation that replaces $({r}_{p},{r}_{q})$ with $({r}_{+},{r}_{})$. Since ${r}_{}$ is pure noise, that is it does not depend on the concentration vector $\mathbf{\mathbf{c}}$ in any way, it has no effect on the mutual information.
We have thus shown that the amount of information that $M$ receptor types contain about the environment when two of the receptors have identical affinity profiles is the same as if there were only $M1$ receptor types. The two redundant receptors can be replaced by a single one with an abundance equal to the sum of the abundances of the two original receptors. The sum of two Gaussian variables with the same mean is Gaussian itself and has a variance equal to the sum of the variances of the two variables, meaning that the noise term ${\eta}_{+}$ in the ${r}_{+}$ response has variance $\frac{{K}_{p}{\sigma}_{p}^{2}+{K}_{q}{\sigma}_{q}^{2}}{{K}_{p}+{K}_{q}}$.
Appendix 3
A nonlinear response example
Estimating the mutual information numerically
Consider an extension of our model in which the responses depend in a nonlinear way on concentrations, but are still subject to pure Gaussian noise:
Note that here we are calculating the average OSN response ${\overline{r}}_{a}={r}_{a}/{K}_{a}$, while in the main text we used the total response ${r}_{a}$. As far as mutual information calculations are concerned, the difference between ${\overline{r}}_{a}$ and ${r}_{a}$ does not matter, as they are related by an invertible transformation.
Unless the functions ${f}_{a}$ are linear, a closedform solution for the mutual information between concentrations and responses cannot be found. It is thus necessary to calculate the mutual information integral numerically. We can still do part of the calculation analytically, though:
In our case, $P(\overline{\mathbf{r}}\mathbf{c})$ is a multivariate Gaussian distribution whose covariance matrix is $\mathrm{\Sigma}{\mathbb{K}}^{1}$ and does not depend on the concentrations. This means that the $\mathbf{\mathbf{c}}$ integral in the second term can be performed independently of the $\overline{\mathbf{\mathbf{r}}}$ integral, in which case it drops out of the calculation, as it is equal to 1. The $\overline{\mathbf{\mathbf{r}}}$ integral is simply the negative entropy of a multivariate Gaussian distribution, and is thus equal to
The first term in Equation (58) is the entropy of the responses, which needs to be calculated numerically. We use a histogram method, in which we split the space of possible responses along each dimension into bins of equal size $\mathrm{\Delta}$. We then estimate the probability in each bin. If ${i}_{1}\mathrm{\dots}{i}_{M}$ indexes the bins, we can then think of the response distribution as a discrete PDF ${P}_{{i}_{1}\mathrm{\dots}{i}_{M}}$, and we can estimate the entropy using
In this approach, the challenge remains to estimate the PDF of the responses,
where $\mathbf{\mathbf{f}}$ is the vector of response functions $\mathbf{\mathbf{f}}=({f}_{1},\mathrm{\dots},{f}_{M})$. We do this using a sampling technique based on the law of large numbers. Given $n$ sample concentration vectors ${c}_{i}$ drawn from the probability distribution $P(\mathbf{\mathbf{c}})$, we have
where $\mathbb{E}}_{P(\mathbf{c})}\{\cdots \$ denotes the expected value under the distribution of concentrations. We use this formula to estimate the histogram elements ${P}_{{i}_{1}\mathrm{\dots}{i}_{M}}$ and then use Equation (60) to estimate the response entropy $H(\overline{\mathbf{\mathbf{r}}})$. We then plug $H(\overline{\mathbf{\mathbf{r}}})$ and Equation (59) into Equation (58) to find the mutual information. Note that we have not assumed anything about the natural distribution of odor concentrations, $P(\mathbf{\mathbf{c}})$, so that we are not restricted to Gaussian environments with this method.
Competitive binding model
The way in which olfactory neurons respond to arbitrary mixtures of odorants is not completely understood. However, simple kinetic models in which different odorant molecules compete for the same receptor binding site have been shown to capture much of the observed behavior (Singh et al., 2018). In such models, the activation of an OSN of type $a$ in response to a set of odorants with concentrations ${c}_{i}$ is given by
where $\mathsf{E}\mathsf{C}\mathsf{50}}_{ai$ is the concentration of odorant $i$ for which the response for the OSN of type $a$ reaches half its maximum, and ${e}_{ai}$ is the maximum response elicited by odorant $i$ in an OSN of type $a$.
Results from a toy problem
The computation time from the method outlined above for calculating mutual information grows exponentially with the dimensionality $M$ of the response space. Additionally, it grows linearly with the number $n$ of samples drawn from the odor distribution, which in turn needs to grow exponentially with the number $N$ of odorants we are considering in order to sample concentration space sufficiently well. For this reason, largescale simulations involving this method are infeasible.
Thus we focused on a simple example with $M=3$ receptors and $N=15$ odorants. We used an arbitrary subset of elements from the fly sensing matrix and a pair of randomlygenerated nonoverlapping environments (Appendix 3—figure 1) to first calculate the optimal receptor distribution using the linear method described in the main text (Appendix 3—figure 2, top). We chose the scale of the environment covariance matrices to get a variability in the responses of around 1, large enough to enter the nonlinear regime when using the nonlinear response function (described below). We then set the total neuron population to ${K}_{\text{tot}}=200$, which put us in an intermediate SNR regime in which all the receptor types were used in the optimal distribution, but their abundances were different (Appendix 3—figure 2, top).
In the linear approximation, we found that receptor 1 is underrepresented in environment 1, while in environment 2 receptor 3 has very low abundance. We wanted to see how much this result is affected by a nonlinear response function. We used a competitive binding model as described above in which the matrix of EC50 values was taken equal to the sensing matrix used in the linear case, and the efficacies ${e}_{ai}$ were all set to 1:
To calculate the mutual information between responses and concentrations for a fixed choice of neuron abundances ${K}_{a}$, we used the procedure outlined above with 20 bins between –0.75 and 1.5 for each of the response dimensions. We sampled $n={10}^{4}$ concentration vectors to build the response histogram. We calculated the information values in both environments at a 10 × 10 grid of OSN abundances (Appendix 3—figure 2, middle row), and found the cell which maximized the information. The OSN abundances at this maximum (Appendix 3—figure 2, bottom) show the same pattern of change as we found in the linear approximation, with receptors 1 and 3 exchanging places as least abundant in the OSN population.
Appendix 4
Random environment matrices
Generating random covariance matrices
Generating plausible olfactory environments is difficult because so little is known about natural odor scenes. However, it is reasonable to expect that there will be some strong correlations. This could, for instance, be due to the fact that an animal’s odor is composed of several different odorants in fixed proportions, and thus the concentrations with which these odorants are encountered will be correlated.
The most straightforward way to generate a random covariance matrix would be to take the product of a random matrix with its transpose, $\mathrm{\Gamma}=M{M}^{T}$. This automatically ensures that the result is positive (semi)definite. The downside of this method is that the resulting correlation matrices tend to cluster close to the identity (assuming that the entries of $M$ are chosen i.i.d.). One way to avoid this would be to use matrices $M$ that have fewer columns than rows, which indeed leads to nontrivial correlations in $\mathrm{\Gamma}$. However, this only generates rankdeficient covariance matrices which means that odorant concentrations are constrained to live on a lowerdimensional hyperplane. This is too strong a constraint from a biological standpoint.
To avoid these shortcomings, we used a different approach for generating random covariance matrices. We split the process into two parts: we first generated a random correlation matrix by the method described below, in which all the variances (i.e. the diagonal elements) were equal to 1; next we multiplied each row and corresponding column by a standard deviation drawn from a lognormal distribution.
In order to generate random correlation matrices, we used a modified form of an algorithm based on partial correlations (Lewandowski et al., 2009). The partial correlation between two variables ${X}_{i}$ and ${X}_{j}$ conditioned on a set of variables $L$ is the correlation coefficient between the residuals ${R}_{i}$ and ${R}_{j}$ obtained by subtracting the best linear fit for ${X}_{i}$ and ${X}_{j}$ using all the variables in $L$. In other words, the partial correlation between ${X}_{i}$ and ${X}_{j}$ is equal to that part of the correlation coefficient that is not explained by the two variables depending on a common set of explanatory variables, $L$. In our case the ${X}_{i}$ are the concentrations of different odorants in the environment and the partial correlations in question are, for example, the correlation between any pair of the odorants conditioned on the remaining ones. We want to construct the unconditioned correlation matrix between the odor concentrations vectors of the environment. There is an algorithm to construct this matrix that starts by randomly drawing the partial correlation between the first two odorants ${X}_{1}$ and ${X}_{2}$ conditioned on the rest, and then recursively reducing the size of the conditioning set while generating more random partial correlations until the unconditioned correlation values are obtained. For details, see Lewandowski et al. (2009).
The specific procedure used in Lewandowski et al. (2009) draws the partial correlation values from beta distributions with parameters depending on the number of elements in the conditioning set $L$. This is done in order to ensure a uniform sampling of correlation matrices. This, however, is not ideal for our purposes because these samples again tend to cluster close to the identity matrix. A simple modification of the algorithm that provides a tunable amount of correlations is to keep the order of the beta distribution fixed $\alpha =\beta =\text{const}$ (see Stack Exchange, at https://stats.stackexchange.com/q/125020). When the parameter $\beta $ is large we obtain environments with little correlation structure, while small $\beta $ values lead to stronger correlations between odorant concentrations. The functions implementing the generation of random environments are available on our GitHub (RRID:SCR_002630) repository at https://github.com/ttesileanu/OlfactoryReceptorDistribution (see environment/generate_random_environment.m and utils/randcorr.m).
Perturbing covariance matrices
When comparing the qualitative results from our model against experiments in which the odor environment changes (IbarraSoria et al., 2017), we used small perturbations of the initial and final environments to estimate error bars on receptor abundances. To generate a perturbed covariance matrix, $\stackrel{~}{\mathrm{\Gamma}}$, from a starting matrix $\mathrm{\Gamma}$, we first took the matrix square root: a symmetric matrix $M$, which obeys
We then perturbed $M$ by adding normallydistributed i.i.d. values to its elements,
and recreated a covariance matrix by multiplying the perturbed square root with its transpose,
This approach ensures that the perturbed matrix $\stackrel{~}{\mathrm{\Gamma}}$ remains a valid covariance matrix—symmetric and positivedefinite—which would not be guaranteed if the random perturbation was added directly to $\mathrm{\Gamma}$. We chose the magnitude $\sigma $ of the perturbation so that the error bars in our simulations are of comparable magnitude to those in the experiments.
We used a similar method for generating the results from Figure 3, where we needed to apply the same perturbation to two different environments. Given the environment covariance matrices ${\mathrm{\Gamma}}_{k}$, with $k\in \{1,2\}$, we took the matrix square root of each environment matrix, ${M}_{k}={\mathrm{\Gamma}}_{k}^{1/2}$. We then added the same perturbation to both, ${\stackrel{~}{M}}_{k}={M}_{k}+P$, then recovered covariance matrices for the perturbed environments by squaring ${\stackrel{~}{M}}_{k}$, ${\stackrel{~}{\mathrm{\Gamma}}}_{k}={\stackrel{~}{M}}_{k}{\stackrel{~}{M}}_{k}^{T}$. In the examples used in the main text, the perturbation $P$ was a matrix in which only one column was nonzero. The elements in this column were chosen from a Gaussian distribution with zero mean and a standard deviation five times larger than the square root of the median element of ${\mathrm{\Gamma}}_{1}$. This choice was arbitrary and was made to obtain a visible change in the optimal receptor abundances between the ‘control’ and ‘exposed’ environments.
Finally, we employed this approach also for generating nonoverlapping environments. Given two environments ${\mathrm{\Gamma}}_{1}$ and ${\mathrm{\Gamma}}_{2}$ and their matrix square roots ${M}_{1}$ and ${M}_{2}$, we reduced the amount of variance in the first half of ${M}_{1}$’s columns and in the second half of ${M}_{2}$’s. We did this by dividing those columns by a constant factor $f$, which in this case we chose to be $f=4$. We then used the resulting matrices ${\stackrel{~}{M}}_{k}$ to generate covariance matrices $\stackrel{~}{\mathrm{\Gamma}}}_{k}={\stackrel{~}{M}}_{k}{\stackrel{~}{M}}_{k}^{T$ with largely nonoverlapping odors.
Appendix 5
Deriving the dynamical model
To turn the maximization requirement into a dynamical model, we employ a gradient ascent argument. Given the current abundances ${K}_{a}$, we demand that they change in proportion to the corresponding components of the information gradient, plus a Lagrange multiplier to impose the constraint on the total number of neurons:
The brain does not have direct access to the overlap matrix $Q$, but it could measure the response covariance matrix $R$ from Equation (13). Thus, we can write the dynamics as
where we used the fact that ${\mathrm{\Sigma}}^{1/2}$ and $\mathbb{K}$ are diagonal and thus commute. These equations do not yet obey the nonnegativity constraint on the receptor abundances. The divergence in the ${K}_{a}^{1}$ term would superficially appear to ensure that positive abundances stay positive, but there is a hidden quadratic divergence in the response covariance term, ${R}_{aa}^{1}$; see Equation (13). To ensure that all constraints are satisfied while avoiding divergences, we multiply the righthandside of Equation (69) by ${K}_{a}^{2}$, yielding
which is the same as Equation (9) from the main text.
If we keep the Lagrange multiplier $\lambda $ constant, the asymptotic value for the total number of neurons ${K}_{\text{tot}}$ will depend on the statistical structure of olfactory scenes. If instead we want to enforce the constraint $\sum {K}_{a}={K}_{\text{tot}}$ for a predetermined ${K}_{\text{tot}}$, we can promote $\lambda $ itself to a dynamical variable,
where $\beta $ is another learning rate. Provided that the dynamics of $\lambda $ is sufficiently slow compared to that of the neuronal populations ${K}_{a}$, this will tune the experienceindependent component of the neuronal death rate until the total population stabilizes at ${K}_{\text{tot}}$.
Appendix 6
Interpretation of diagonal elements of the inverse overlap matrix
In the main text we saw that the diagonal elements of the inverse overlap matrix ${Q}_{aa}^{1}$ were related to the abundances of OSNs ${K}_{a}$. Specifically,
where $\lambda $ is a Lagrange multiplier imposing the constraint on the total number of neurons. As noted around Equation (13) above, the overlap matrix $Q$ is related to the response covariance matrix $R$: in particular, $Q$ is equal to $R$ when there is a single receptor of each type (${K}_{a}=1$) and there is no noise (${\sigma}_{a}=0$). That is, the overlap matrix measures the covariances between responses in the absence of noise. This means that its inverse $A={Q}^{1}$ is effectively a socalled ‘precision matrix’. Diagonal elements of a precision matrix are inversely related to corresponding diagonal elements of the covariance matrix (i.e. the variances), but, as we will see below, they are also monotonically related to parameters that measure how well each receptor response can be linearly predicted from all the others. Since receptor responses that either do not fluctuate much or whose values can be guessed based on the responses of other receptors are not very informative, we would expect that abundances ${K}_{a}$ are low when the corresponding diagonal elements of the inverse overlap matrix ${A}_{aa}$ are high, which is what we see. In the following we give a short derivation of the connection between the diagonal elements of precision matrices and linear prediction of receptor responses.
Let us work in the particular case in which there is one copy of each receptor and where there is no noise, so that $Q=R$, that is ${Q}_{ij}=\u27e8{r}_{i}{r}_{j}\u27e9\u27e8{r}_{i}\u27e9\u27e8{r}_{j}\u27e9$. Without loss of generality, we focus on calculating the first diagonal element of the inverse overlap matrix, ${A}_{11}$, where $A={Q}^{1}$. For notational convenience, we will also denote the meancentered first response variable by $y\equiv {r}_{1}\u27e8{r}_{1}\u27e9$, and the subsequent ones by ${x}_{a}\equiv {r}_{a+1}\u27e8{r}_{a+1}\u27e9$. Then the covariance matrix $Q$ can be written in block form
where $M$ is
and $\mathbf{\mathbf{x}}$ is a column vector containing the ${x}_{a}$ variables. Using the definition of the inverse together with Laplace’s formula for determinants, we get
Using the Schur determinant identity (derived above) on the block form (Equation (73)) of the matrix $Q$,
where we used the fact that the argument of the second determinant is a scalar.
Now, consider approximating the first response variable $y$ by a linear function of all the others:
where $q$ is the residual. Note that we do not need an intercept term because we meancentered our variables, $\u27e8y\u27e9=\u27e8x\u27e9=0$. Finding the coefficients $\mathbf{\mathbf{a}}$ that lead to a best fit (in the leastsquares sense) requires minimizing the variance of the residual, and a short calculation yields
where $M$ is the same as the matrix defined in Equation (74).
The coefficient of determination ${\rho}^{2}$ is defined as the ratio of explained variance to total variance of the variable $y$,
Comparing this to Equation (76), we see that
showing that the diagonal elements of the precision matrix are monotonically related to the goodnessoffit parameter ${\rho}^{2}$ that indicates how well the corresponding variable can be linearly predicted by all the other variables. In addition, the inverse dependence on the variance of the response ${\u27e8y\u27e9}^{2}$ shows that variables that do not fluctuate much (low ${\u27e8y\u27e9}^{2}$) lead to high diagonal values of the precision matrix . From Equation (72), we see that these variances should be considered ‘large’ or ”small’ in comparison with the noise level in each receptor (${\sigma}_{a}$). Since receptor responses that either do not fluctuate much or whose values can be guessed based on the responses of other receptors are not very informative, we should find that receptor abundances ${K}_{a}$ are low when the corresponding diagonal elements of the inverse overlap matrix ${A}_{aa}={Q}_{aa}^{1}$ are high.
Data availability
All the code necessary to reproduce our results and the figures from the paper is available on GitHub, at https://github.com/ttesileanu/OlfactoryReceptorDistribution (copy archived at https://github.com/elifesciencespublications/OlfactoryReceptorDistribution). The olfactory receptor affinity data were originally published in Hallem et al. (2006) and Saito et al. (2009), and the olfactory receptor expression levels in mouse were originally published in IbarraSoria et al. (2017).
References

Towards a theory of early visual processingNeural Computation 2:308–320.https://doi.org/10.1162/neco.1990.2.3.308

Sensory Communication217–234, Possible principles underlying the transformations of sensory messages, Sensory Communication, MIT Press.

Postnatal odorant exposure induces peripheral olfactory plasticity at the cellular levelJournal of Neuroscience 34:4857–4870.https://doi.org/10.1523/JNEUROSCI.068813.2014

Neurogenesis and cell death in olfactory epitheliumJournal of Neurobiology 30:67–81.https://doi.org/10.1002/(SICI)10974695(199605)30:1<67::AIDNEU7>3.0.CO;2E

Design of a trichromatic cone arrayPLoS Computational Biology 6:e1000677.https://doi.org/10.1371/journal.pcbi.1000677

Comparative morphometry of the nasal cavity in rats and miceJournal of Anatomy 135:83–88.

Mammalian Brains Are Made of These: A Dataset of the Numbers and Densities of Neuronal and Nonneuronal Cells in the Brain of Glires, Primates, Scandentia, Eulipotyphlans, Afrotherians and Artiodactyls, and Their Relationship with Body MassBrain, Behavior and Evolution 86:145–163.https://doi.org/10.1159/000437413

Mechanisms of olfactory discrimination: converging evidence for common principles across phylaAnnual Review of Neuroscience 20:595–631.https://doi.org/10.1146/annurev.neuro.20.1.595

In search of the structure of human olfactory spaceFrontiers in Systems Neuroscience 5:1–8.https://doi.org/10.3389/fnsys.2011.00065

A simple coding procedure enhances a neuron's Information CapacityZeitschrift Für Naturforschung C 36:910–912.https://doi.org/10.1515/znc198191040

Generating random correlation matrices based on vines and extended onion methodJournal of Multivariate Analysis 100:1989–2001.https://doi.org/10.1016/j.jmva.2009.04.008

Scaling of mammalian ethmoid bones can predict olfactory organ size and performanceProceedings of the Royal Society B: Biological Sciences 272:957–962.https://doi.org/10.1098/rspb.2004.2993

Novel Behavioral Paradigm Reveals Lower Temporal Limits on Mouse Olfactory DecisionsThe Journal of Neuroscience 35:11667–11673.https://doi.org/10.1523/JNEUROSCI.469314.2015

The Value of Information for Populations in Varying EnvironmentsJournal of Statistical Physics 142:1124–1166.https://doi.org/10.1007/s1095501101662

BookIdentified Glomeruli in the Antennal Lobes of Insects: In Variance, Sexual Variation and Postembryonic DevelopmentIn: Singh R. N, Strausfeld N. J, editors. Neurobiology of Sensory Systems. Boston, MA: Springer US. pp. 355–375.

Odor coding by a Mammalian receptor repertoireScience Signaling 2:ra9.https://doi.org/10.1126/scisignal.2000016

Optimal Prediction in the Retina and Natural Motion StatisticsJournal of Statistical Physics 162:1309–1323.https://doi.org/10.1007/s109550151439y

Olfactory sensory neurons are trophically dependent on the olfactory bulb for their prolonged survivalThe Journal of Neuroscience 12:3896–3919.https://doi.org/10.1523/JNEUROSCI.121003896.1992

Natural image statistics and neural representationAnnual Review of Neuroscience 24:1193–1216.https://doi.org/10.1146/annurev.neuro.24.1.1193

Nasal morphometry in marmosets: loss and redistribution of olfactory surface areaThe Anatomical Record 297:2093–2104.https://doi.org/10.1002/ar.23029

Predicting odor perceptual similarity from odor structurePLoS Computational Biology 9:e1003184.https://doi.org/10.1371/journal.pcbi.1003184

Predictive coding: a fresh view of inhibition in the retinaProceedings of the Royal Society of London. Series B, Biological sciences 216:427–459.https://doi.org/10.1098/rspb.1982.0085

A theory of maximizing sensory informationBiological Cybernetics 68:23–29.https://doi.org/10.1007/BF00203134

Theoretical predictions of spatiotemporal receptive fields of fly LMCs, and experimental validationJournal of Comparative Physiology A 171:157–170.https://doi.org/10.1007/BF00188924

Spatiotemporal contrast sensitivity of early visionVision Research 33:257–267.https://doi.org/10.1016/00426989(93)90163Q

Independent component filters of natural images compared with simple cells in primary visual cortexProceedings of the Royal Society of London. Series B: Biological Sciences 265:359–366.https://doi.org/10.1098/rspb.1998.0303

A Robust Feedforward Model of the Olfactory SystemPLOS Computational Biology 12:e1004850.https://doi.org/10.1371/journal.pcbi.1004850
Article and author information
Author details
Funding
Simons Foundation (400425)
 Vijay Balasubramanian
Aspen Center for Physics (PHY160761)
 Vijay Balasubramanian
Swartz Foundation
 Tiberiu Teşileanu
National Science Foundation (PHY1734030)
 Tiberiu Teşileanu
 Vijay Balasubramanian
United States  Israel Binational Science Foundation (2011058)
 Vijay Balasubramanian
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
Acknowledgements
We thank Joel Mainland and David Zwicker for helpful discussions, and Elissa Hallem, Joel Mainland, and Darren Logan for olfactory receptor affinity data. This work was supported by a grant from the Simons Foundation/SFARI Mathematical Modeling in Living Systems program (400425, VB). VB was also supported by Aspen Center for Physics NSF grant PHY160761 and US–Israel Binational Science Foundation grant 2011058. TT was supported by the Swartz Foundation. This work was also supported by NSF grant PHY1734030 (Center for the Physics of Biological Function).
Copyright
© 2019, Teşileanu et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics

 4,706
 views

 458
 downloads

 24
 citations
Views, downloads and citations are aggregated across all versions of this paper published by eLife.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading

 Computational and Systems Biology
 Physics of Living Systems
Planar cell polarity (PCP) – tissuescale alignment of the direction of asymmetric localization of proteins at the cellcell interface – is essential for embryonic development and physiological functions. Abnormalities in PCP can result in developmental imperfections, including neural tube closure defects and misaligned hair follicles. Decoding the mechanisms responsible for PCP establishment and maintenance remains a fundamental open question. While the roles of various molecules – broadly classified into “global” and “local” modules – have been wellstudied, their necessity and sufficiency in explaining PCP and connecting their perturbations to experimentally observed patterns have not been examined. Here, we develop a minimal model that captures the proposed features of PCP establishment – a global tissuelevel gradient and local asymmetric distribution of protein complexes. The proposed model suggests that while polarity can emerge without a gradient, the gradient not only acts as a global cue but also increases the robustness of PCP against stochastic perturbations. We also recapitulated and quantified the experimentally observed features of swirling patterns and domineering nonautonomy, using only three free model parameters  the rate of protein binding to membrane, the concentration of PCP proteins, and the gradient steepness. We explain how selfstabilizing asymmetric protein localizations in the presence of tissuelevel gradient can lead to robust PCP patterns and reveal minimal design principles for a polarized system.

 Computational and Systems Biology
 Physics of Living Systems
Synthetic genetic oscillators can serve as internal clocks within engineered cells to program periodic expression. However, celltocell variability introduces a dispersion in the characteristics of these clocks that drives the population to complete desynchronization. Here, we introduce the optorepressilator, an optically controllable genetic clock that combines the repressilator, a threenode synthetic network in E. coli, with an optogenetic module enabling to reset, delay, or advance its phase using optical inputs. We demonstrate that a population of optorepressilators can be synchronized by transient green light exposure or entrained to oscillate indefinitely by a train of short pulses, through a mechanism reminiscent of natural circadian clocks. Furthermore, we investigate the system’s response to detuned external stimuli observing multiple regimes of global synchronization. Integrating experiments and mathematical modeling, we show that the entrainment mechanism is robust and can be understood quantitatively from single cell to population level.