Frontend WeberFechner gain control enhances the fidelity of combinatorial odor coding
Abstract
We showed previously (GorurShandilya et al., 2017) that Drosophila olfactory receptor neurons (ORNs) expressing the coreceptor Orco scale their gain inversely with mean odor intensity according to WeberFechner’s law. Here, we show that this frontend adaptation promotes the reconstruction of odor identity from dynamic odor signals, even in the presence of confounding background odors and rapid intensity fluctuations. These enhancements are further aided by known downstream transformations in the antennal lobe and mushroom body. Our results, which are applicable to various odor classification and reconstruction schemes, stem from the fact that this adaptation mechanism is not intrinsic to the identity of the receptor involved. Instead, a feedback mechanism adjusts receptor sensitivity based on the activity of the receptorOrco complex, according to WeberFechner’s law. Thus, a common scaling of the gain across Orcoexpressing ORNs may be a key feature of ORN adaptation that helps preserve combinatorial odor codes in naturalistic landscapes.
https://doi.org/10.7554/eLife.45293.001Introduction
Animals identify and discriminate odors using olfactory receptors (Ors) expressed in olfactory receptor neurons (ORNs) (Joseph and Carlson, 2015; Buck and Axel, 1991; Clyne et al., 1999; Vosshall et al., 1999). Individual ORNs, which typically express a single Or, respond to many odorants, while individual odorants activate many distinct ORNs (Friedrich and Korsching, 1997; Hallem and Carlson, 2006; Wang et al., 2010; Nara et al., 2011). Odors are thus encoded by the combinatorial patterns of activity they elicit in the sensing periphery (Malnic et al., 1999; Wang et al., 2010; Hildebrand and Shepherd, 1997; Hallem and Carlson, 2006; de Bruyne et al., 2001; Friedrich and Korsching, 1997), and these patterns are decoded downstream into behavioral response (Wilson, 2013; Davies et al., 2015). Still, ethologically relevant odors are often mixed with background ones (Saha et al., 2013; Renou et al., 2015) and intensity can vary widely and rapidly as odors are carried by the wind (Murlis et al., 1992; Weissburg, 2000; Celani et al., 2014; Cardé and Willis, 2008). How are odors recognized reliably despite these confounds? In Drosophila melanogaster, ORN dose response curves exhibit similar Hill coefficients but distinct powerlaw distributed activation thresholds (Hallem and Carlson, 2006; Si et al., 2019), which together with inhibitory odorants enhance coding capacity (Si et al., 2019; Cao et al., 2017; Hallem and Carlson, 2006; Stevens, 2016). In antennal lobe (AL) glomeruli, mutual lateral inhibition normalizes population response, reducing the dependency of activity patterns on odor concentration (Asahina et al., 2009; Olsen et al., 2010). Further downstream, sparse connectivity to the mushroom body (MB) helps maintain neural representations of odors, and facilitates compressed sensing and associative learning schemes (Caron et al., 2013; LitwinKumar et al., 2017; Krishnamurthy et al., 2017; Dasgupta et al., 2017). Finally, temporal features of neural responses contribute to concentrationinvariant representations of odor identity (Brown et al., 2005; Raman et al., 2010; Gupta and Stopfer, 2014; Wilson et al., 2017).
Here, we examine how shorttime ORN adaptation at the very frontend of the insect olfactory circuit contributes to the fidelity of odor encoding. Our theoretical study is motivated by the recent discovery of invariances in the signal transduction and adaptation dynamics of ORNs expressing the coreceptor Orco. ORN response is initiated upon binding of odorant molecules to olfactory receptors (ORs), opening the ion channels they form with the coreceptor Orco (Larsson et al., 2004; Butterwick et al., 2018). Because of differences in odorreceptor affinities, the responses of ORNs to diverse odorants of the same concentration differ widely (Hallem and Carlson, 2006; Montague et al., 2011; Stensmyr et al., 2012). In contrast, downstream from this input nonlinearity, signal transduction and adaptation dynamics exhibit a surprising degree of invariance with respect to odorreceptor identity: reversecorrelation analysis of ORN response to fluctuating stimuli produces highly stereotyped, concentrationinvariant response filters (Martelli et al., 2013; Si et al., 2019; GorurShandilya et al., 2017).
These properties stem in part from an apparently invariant adaptive scaling law in ORNs: gain varies inversely with mean odor concentration according to the WeberFechner Law of psychophysics (Weber, 1996; Fechner, 1860), irrespective of the odorreceptor combination (GorurShandilya et al., 2017; Cafaro, 2016; Cao et al., 2016). This invariance can be traced back to adaptative feedback mechanisms in odor transduction, upstream of ORN firing (Nagel and Wilson, 2011; Cao et al., 2016; Cafaro, 2016; GorurShandilya et al., 2017), which depend on the activity of the signaling pathway rather than on the identity of its receptor (Nagel and Wilson, 2011). The generality of the adaptive scaling suggests it could be mediated by the highly conserved Orco coreceptor (Butterwick et al., 2018; Getahun et al., 2013; Getahun et al., 2016; Guo et al., 2017), which has been already been implicated in other types of odor adaptation, taking place over longer timescales (Guo and Smith, 2017; Guo et al., 2017).
While in a simpler system such as E. coli chemotaxis (Waite et al., 2018), adaptive feedback via the WeberFechner Law robustly maintains sensitivity over concentration changes, the implication for a multiplechannel system – which combines information from hundreds of cells with overlapping receptive fields – is less clear. Here, we combine a biophysical model of ORN adaptive response and neural firing with various sparse signal decoding frameworks to explore how ORN adaptation with WeberFechner scaling affects combinatorial coding and decoding of odor signals spanning varying degrees of intensity, molecular complexity, and temporal structure. We find that this frontend adaptive mechanism promotes the accurate discrimination of odor signals from backgrounds of varying molecular complexity, and aids other known mechanisms of neural processing in the olfactory circuit to maintain representations of odor identity across environmental changes.
Results
Model of ORN sensing repertoire
To model ORN firing rates in response to timedependent odor signals, we extended a minimal model (GorurShandilya et al., 2017) that reproduces the WeberFechner gain adaptation and firing rate dynamics measured in individual Drosophila ORNs in response to Gaussian and naturalistic signals (code available on GitHub, Kadakia, 2019; copy archived at https://github.com/elifesciencespublications/ORNWLgaincontrol).
We consider a repertoire of $M=50$ ORN types that each express one type of Or together with the coreceptor Orco (Larsson et al., 2004). Within ORNs of type $a=1,\mathrm{\dots},M$, OrOrco complexes form nonselective cation channels (Butterwick et al., 2018) (Figure 1A) that switch between active and inactive conformations, while simultaneously binding to odorants $i$ with affinity constants, ${K}_{ai}^{*}$ and ${K}_{ai}$, respectively (Nagel and Wilson, 2011; GorurShandilya et al., 2017). For simplicity, we only consider agonists, that is ${K}_{ai}^{*}>{K}_{ai}$, and assume receptors can only bind one odorant at a time. The analysis can easily be extended to include inhibitory odorants, which increases coding capacity (Cao et al., 2017). Dissociation (inverse affinity) constants are chosen from a power law distribution ($\alpha =0.35$) recently found across ORNodor pairs in Drosophila larvae (Si et al., 2019). For a handful of ORNs, we choose a very large value for one of the ${K}_{ai}^{*}$ to mimic high responders to private odorants relevant to innate responses (Stensmyr et al., 2012). These private odors do not affect the general findings.
Assuming that odorant binding and conformation changes are faster than other reactions in the signaling pathway, the fraction of channels of type $a$ that are active at steady state is:
$\mathrm{C}}_{a$ and $\mathrm{C}}_{a}^{\ast$ represent unbound channels in the inactive and active conformation. Here, ${\mathbf{\mathbf{K}}}_{a}\cdot \mathbf{\mathbf{s}}(t)={\sum}_{i}^{N}{K}_{ai}{s}_{i}(t)$, where ${s}_{i}(t)$ is the timedependent concentration of the $i$th monomolecular component of the odor signal $\mathbf{\mathbf{s}}(t)$ at time $t$ (Figure 1B). $N=150$ is the size of the molecular odorant space (Figure 1B). Equation 1 can be rearranged as (derivation in Materials and methods):
The two terms in the exponential represent the change in the channel’s free energy due to the binding of odorant $i$, and the free energy difference ${\u03f5}_{a}$ between the unbound states $\mathrm{C}}_{a$ and $\mathrm{C}}_{a}^{\ast$, in units of ${k}_{B}T$. Because ${K}_{ai}^{*}>{K}_{ai}$, a sudden increase in the concentration of excitatory odor results in an increase in activity ${A}_{a}$.
Upon prolonged stimulation, ORNs adapt. At least one form of adaptation, which takes place over short time scale, $\tau \simeq 250\phantom{\rule{thinmathspace}{0ex}}\phantom{\rule{thinmathspace}{0ex}}\mathrm{m}\mathrm{s}$ (GorurShandilya et al., 2017), involves a negative feedback of the OrOrco channel activity onto the channel sensitivity (Nagel and Wilson, 2011; GorurShandilya et al., 2017). To model this adaptation process, we assume that inward currents elicited by activating OrOrco channels eventually result in an increase of the free energy difference ${\u03f5}_{a}(t)$, possibly via a feedback onto Orco (Butterwick et al., 2018):
where $\u03f5}_{\mathrm{L},a}<{\u03f5}_{a}(t)<{\u03f5}_{\mathrm{H},a$. The lower bound $\u03f5}_{\mathrm{L},a$ determines the spontaneous activity of the channel. The higher bound $\u03f5}_{\mathrm{H},a$ determines the concentrations of odors at which adaptation is unable to keep up and saturation occurs (GorurShandilya et al., 2017). Through these dynamics, ${\u03f5}_{a}(t)$ can compensate for changes in free energy due to ligand binding (see Equation 2), returning the activity ${A}_{a}$ towards an adapted level ${A}_{0a}$ above the spontaneous activity. Since ${\u03f5}_{a}$ is bounded below, a minimum amount of signal intensity is needed for adaptation to kick in. Finally, the firing rate is modeled by passing the activity ${A}_{a}(t)$ through the derivativetaking bilobed filter $h(t)$ and a rectifying nonlinearity $f$ (GorurShandilya et al., 2017):
where $\otimes $ is convolution. When deconvolved from stimulus dynamics, the shapes of the temporal kernels of Drosophila ORNs that express Orco tend to be stereotyped for many odorreceptor combination (Martelli et al., 2013; GorurShandilya et al., 2017; Si et al., 2019) (although there are known exceptions such as supersustained responses Montague et al., 2011). Moreover, adaptation is not intrinsic to the receptor (Nagel and Wilson, 2011). Accordingly, for simplicity $\tau $, $h(t)$, and $f$ are assumed independent of receptor and odorant identities.
This minimal model reproduces the essential features of ORN response to odorant pulses (Nagel and Wilson, 2011; Martelli et al., 2013; Cao et al., 2016). In the absence of stimulus, ORNs fire spontaneously at rates (1–10 Hz) (Hallem and Carlson, 2006) set by the lower free energy bound ${\u03f5}_{\text{L},a}$, which we choose from a normal distribution (Figure 1D). For sufficiently strong stimuli, adaptation causes ${\u03f5}_{a}$ to increase, compensating for the drop in free energy difference due to ligand binding. This gradually reduces the firing rate to a steady state level $r({A}_{0a})\simeq $ 30–40 Hz (GorurShandilya et al., 2017) (Figure 1D). The diversity of temporal firing responses and tuning curves measured experimentally (Hallem and Carlson, 2006; Montague et al., 2011; Brown et al., 2005; Gupta and Stopfer, 2014; Raman et al., 2010) arise naturally in the model due to the distribution of chemical affinity constants and the nonlinearity of Equation 2 (Figure 1BFigure 1E).
The model also reproduces WeberFechner scaling of the gain with the inverse of the mean odorant intensity ${\overline{s}}_{i}$(GorurShandilya et al., 2017; Cao et al., 2016). For small fluctuations $\mathrm{\Delta}{s}_{i}$ around ${\overline{s}}_{i}$, we have from Equation 2 that $\mathrm{\Delta}{A}_{a}/\mathrm{\Delta}{s}_{i}\simeq {A}_{a}\left({\overline{s}}_{i}\right)\left(1{A}_{a}\left({\overline{s}}_{i}\right)\right)/{\overline{s}}_{i}$, whereby Weber’s Law is satisfied provided ${A}_{a}({\overline{s}}_{i})$ is approximately constant (derivation in Materials and methods). In our model, since the rate of adaptation depends only on the activity of the ion channel (right handside of Equation 3), then in the adapted state we have ${A}_{a}\left({\overline{s}}_{i}\right)\simeq {A}_{0a}$, ensuring that the gain scales like $1/{\overline{s}}_{i}$. This process adjusts the sensitivity of the ORN by matching the dose responses to the mean signal concentration, while maintaining their logslopes (Figure 1F). However, for foreground odors mixed with background odors to which the system has adapted, the dose response curves now exhibit backgrounddependent shifts (Figure 1G).
While this phenomenological model could be extended to include further details – for example, we could relax the quasisteadystate assumption in Equation 2, use a more complex model for channel adaptation and neural firing (GorurShandilya et al., 2017), or consider feedforward mechanisms in addition to negative integral feedback (Schulze et al., 2015) – this minimally parameterized form captures the key dynamical properties of Orcoexpressing ORNs relevant to our study: receptorindependent adaptation (Nagel and Wilson, 2011) with WeberFechner scaling (GorurShandilya et al., 2017; Cafaro, 2016; Cao et al., 2016) that maintains response time independent of mean stimulus intensity (Martelli et al., 2013; GorurShandilya et al., 2017), along with a diversity of temporal firing patterns in response to a panel of monomolecular odorants (Hallem and Carlson, 2006; Montague et al., 2011; Brown et al., 2005; Gupta and Stopfer, 2014; Raman et al., 2010) (Figure 1D–1E).
Frontend WeberFechner adaptation preserves odor coding among background and intensity confounds
The identity of an odor is encoded by the pattern of ORN firing responses. However, when a novel foreground odor is presented atop an existing background odor, this pattern may depend also on the background odor, rendering ORN responses less informative about foreground odor identity. To understand how frontend WeberFechner adaptation might help encode novel foreground odors in the presence of background odors, we considered environments containing various combinations of foreground odors $\mathbf{\mathbf{s}}$ and background odors $\overline{\mathbf{\mathbf{s}}}$, and asked how similar are the ORN responses $\mathbf{\mathbf{r}}$ to a given $\mathbf{\mathbf{s}}$ but different $\overline{\mathbf{\mathbf{s}}}$.
Since it is not possible to visualize the 50dimensional space of ORN responses, we projected ORN responses onto a twodimensional space using tdistributed stochastic neighbor embedding (tSNE) (van der Maaten and Hinton, 2008). Like principle component analysis (PCA), tSNE allows a visualization of highdimensional objects in such a way that desirable features of the original dataset are preserved (Figure 2A). PCA, for example, retains much of the data variance. tSNE retains the proximity of an object to its nearest neighbors. Specifically, it constructs a probability distribution $Q}_{\mathrm{H}$ based on pairwise distances between nearby objects, assigning higher probability to closer objects. It then determines where the objects would live in a lower dimensional space, such that the analogous distribution $Q}_{\mathrm{L}$ in this space is most similar to $Q}_{\mathrm{H}$. tSNE is widely used to cluster objects (in our case, ORN responses $\mathbf{\mathbf{r}}$ to different foreground odors on top of diverse background odors) by similarity (here, foreground odor identity). However, because tSNE uses local information from only nearest neighbors, global distances and scales are not preserved (Zhou and Sharpee, 2018). Thus, we use tSNE only for visualization. To more rigorously quantify how foreground identity is preserved in ORN activity, we calculate the mutual information (MI) between foreground odor $\mathbf{\mathbf{s}}$ and ORN firing rates $\mathbf{\mathbf{r}}$ in the 50dimensional space (Materials and methods). The MI quantifies how much information a response contains about the stimulus. High MI means that responses exhibit larger variability for different stimuli than for repeated presentations of the same stimulus. In our case, this would be true if $\mathbf{\mathbf{r}}$ were uniquely defined for different foregrounds $\mathbf{\mathbf{s}}$, irrespective of the background $\overline{\mathbf{\mathbf{s}}}$. Conversely, the MI would be low if responses varied more by background $\overline{\mathbf{\mathbf{s}}}$ than by foreground $\mathbf{\mathbf{s}}$.
We first examined how an adaptive or nonadaptive ORN repertoire encodes odor identity in an odor environment that contains a foreground odor $\mathbf{\mathbf{s}}$ atop a background odor $\overline{\mathbf{\mathbf{s}}}$ (Figure 2B). Both odors are sparse mixtures, with $K\ll N$ odorants of similar concentrations, odor ‘identity’ being the particular set of odorants in the mixture. In the adaptive case, we assume that the system has fully adapted to the background $\overline{\mathbf{\mathbf{s}}}$ before the foreground $\mathbf{\mathbf{s}}$ is presented. This is enacted by calculating the firing response to the foreground odor $\mathbf{\mathbf{r}}(\mathbf{\mathbf{s}})$ only after having set the ${\u03f5}_{a}$ in Equation 2 to their steady state values in response to the background odor $\overline{\mathbf{\mathbf{s}}}$:
where we have introduced the new parameter ${\beta}_{a}$ to allow us to control the scaling of gain adaptation: for ${\beta}_{a}=0$ the system exactly follows WeberFechner’s law, while for ${\beta}_{a}=1$ there is no adaptation. For small but nonzero ${\beta}_{a}$, the inverse gain scales sublinearly (see Materials and methods), and the adapted activity ${A}_{a}(\overline{\mathbf{\mathbf{s}}})$ increases weakly with background $\overline{\mathbf{\mathbf{s}}}$. In experiments, small deviations from the strict WeberFechner scaling on the order of $\beta \simeq 0.1$ are observed (see extended figures in GorurShandilya et al., 2017).
With WeberFechner’s law in place for all ORNs (${\beta}_{a}=0$) responses cluster by the identity of foreground odor, showing that the repertoire of ORNs appropriately encodes the identity of novel odors irrespective of background signals – once these backgrounds have been ‘adapted away’ (Figure 2B). This is the case regardless of whether ${A}_{0a}$ is identical or different across neurons (Figure 2—figure supplement 1). In contrast, when the system is nonadaptive, (${\beta}_{a}=1$), the responses exhibit weaker separations by odor identity (Figure 2B). Similarly, responses across different odor intensities are well separated by odor identity in the adaptive system, but less so in the nonadaptive system (Figure 2C). Calculating the mutual information between odor and ORN response in time shows that the adaptive system retains coding capacity as it confronts novel odors (Figure 2—figure supplement 2), whereas the nonadaptive system maintains coding capacity in a far more limited range of odor concentration.
To what extent do the benefits of frontend adaptation for odor coding depend on the precise WeberFechner scaling? We repeated the analysis from Figure 2B for increasing values of ${\beta}_{a}=\beta $ between zero (Weber’s law) (perfect adaptation) and one (no adaptation). To generalize Figure 2B, we now let the intensities range over two decades. As $\beta $ increases, the capacity of the system to cluster responses by odor identity degrades (Figure 2D). Introducing diversity among ORNs by distributing ${\beta}_{a}$’s uniformly between 0 and $2\beta $ (so that the mean is $\beta $) slightly increases performance at high $\beta $ but reduces it at low $\beta $ (Figure 2D). Overall, performance of odor coding degrades with $\beta $, as poorly adapting ORNs begin to saturate (Figure 2D).
Interestingly, besides this general trend, we find that for $\beta $ very close to zero, a small deviation from WeberFechner’s law instead improves odor coding. This arises because of the nonlinearity in the onset of adaptation: adaptation kicks in only when the strength of stimulus is sufficient for the response ${A}_{a}$ to exceed ${A}_{0a}$, so that the right handside of Equation 3 is positive. The minimum background intensity $\overline{s}$ required for this to happen is given by ${\u03f5}_{\mathrm{L},a}={\u03f5}_{a}(\overline{s})$, which, according to equation Equation 5, increases with $\beta $. This initial effect increases odor coding performance, as the firing rates can distribute more broadly across the dynamical range of the ORNs, before adaptation is effected (Figure 2E). Note that this effect is not specific to our model. A similar enhancement would be observed if Weber’s Law were maintained, but kicked in only above a minimum signal intensity. Thus, while WeberFechner scaling largely preserves the representation of foreground odor identity amid backgrounds, in some cases it may benefit from a slight relaxation so that the full dynamical range of the ORNs can be exploited.
Frontend adaptation enhances odor decoding in complex environments
Given that frontend adaptation helps maintain combinatorial odor codes in the presence of backgrounds, we wondered how it affects the capability to decode odor signals from ORN response. One potentially complicating factor is the disparity between sensor dimension and stimulus dimension: while Drosophila only express ∼60 Or genes (Vosshall et al., 2000), the space of odorants is far greater (Krishnamurthy et al., 2017). An $N$dimensional odor signal would naively need $N$ sensory neurons to decode it – one for each odorant. However, naturally occurring odors are sparse, typically comprised of only a few odorants. Enforcing sparsity of the signal during decoding greatly restricts the number of possible odors consistent with a given ORN response, suggesting that such highdimensional signals might be inferred from less than $N$ ORNs. Indeed, the decoding of sufficiently sparse signals from lower dimensional responses is rigorously guaranteed by the theory of compressed sensing (CS) (Donoho, 2006; Candès et al., 2006). It is unknown whether CS is implemented in the Drosophila olfactory circuit (Pehlevan et al., 2017). Here, we use this framework mainly as a tool to quantify how frontend adaptation potentially affects odor decoding, later verifying our conclusions with other classification techniques that incorporate the known architecture of the olfactory system.
CS is performed as a constrained linear optimization. The constraints in the optimization are $\mathbf{\mathbf{r}}=\mathbf{\mathbf{D}\mathbf{s}}$, where $\mathbf{\mathbf{s}}$ is the stimulus to be estimated, $\mathbf{\mathbf{D}}$ is the response matrix, and $\mathbf{\mathbf{r}}$ is the vector of ORN responses. The cost function to be minimized, $C={\sum}_{i}{s}_{i}$, enforces sparsity by driving the estimate of each odorant component to zero; the constraints balance this tendency by simultaneously enforcing information from the ORN firing rates. The result is a reconstructed odor signal $\widehat{\mathbf{\mathbf{s}}}$ that is as sparse as possible, consistent with the ORN responses. In practice, one uses a linear optimization routine to numerically minimize ${\sum}_{i}{s}_{i}$ over ${s}_{i}$, subject to $\mathbf{\mathbf{r}}=\mathbf{\mathbf{D}\mathbf{s}}$. The result is an estimate of the magnitude of each signal component ${s}_{i}$. Thus, both the identity and the intensity of the odor signal are estimated.
To incorporate this linear framework of CS into our nonlinear odor encoding model, we treat the nonlinear odor encoding exactly, but approximate the decoding to first order around the background concentration (Figure 3A). Specifically, we use Equations 24 to generate ORN responses $\mathbf{\mathbf{r}}$ for sparse odors $\mathbf{\mathbf{s}}$ having $K\ll N$ nonzero components ${s}_{i}={\overline{s}}_{i}+\mathrm{\Delta}{s}_{i}$, where the mean concentration is ${\overline{s}}_{i}$. To estimate signals using CS, we minimize ${\sum}_{i}\mathrm{\Delta}{s}_{i}$ while enforcing the constraints $\mathbf{\mathbf{r}}=\mathbf{\mathbf{D}}\mathrm{\Delta}\mathbf{\mathbf{s}}$, where $\mathbf{\mathbf{D}}$ is the linearization of Equation 2 around ${\overline{s}}_{i}$ (details in Materials and methods). The perturbations are chosen as $\mathrm{\Delta}{s}_{i}\sim \mathcal{N}({s}_{0}/3,{s}_{0}/9)$, where ${\overline{s}}_{i}={s}_{0}$. This linearization simplifies the CS decoding – namely it enforces a single, global minimum – but it is not critical for our general results; see Materials and methods and Figure 3—figure supplement 5. We perform the minimization using the sequential least squares algorithm, producing an estimate of the concentration $\mathrm{\Delta}{s}_{i}$ of each individual odorant. The matrix $\mathbf{\mathbf{D}}$ depends on ${\u03f5}_{a}$, and as above, we assume precise adaptation by setting ${\u03f5}_{a}$ to their steady state values in response to the background odor alone (via Equation 5 with $\beta =0$). In the nonadaptive case, ${\u03f5}_{a}$ are held at their minimum values $\u03f5}_{\mathrm{L},a$.
We first examine how foreground odors are recognized when mixed with background odors of a distinct identity but similar intensities, quantifying decoding accuracy as the percentage of odors correctly decoded within some tolerance (see Materials and methods). Without adaptation, accuracy is maintained within the range of receptor sensitivity for monomolecular backgrounds but is virtually eliminated as background complexity rises (Figure 3B). The range of sensitivity is broader in the adaptive system and is substantially more robust across odor concentration and complexity (Figure 3C).
In realistic odor environments, the concentration and duration of individual odor whiffs vary widely (Celani et al., 2014). We wondered how well a frontend adaptation mechanism with a single timescale $\tau $ could promote odor identity detection in such environments. As inputs to our coding/decoding framework, we apply a naturalistic stimulus intensity recorded from a photoionization detector (GorurShandilya et al., 2017) (Figure 3D), to which we randomly assign sparse identities from the $N$dimensional odorant space (odor concentration fluctuates in time, but identity is fixed). To mimic background confounds, we combine these signals with a static odor background of a different identity. We decode the odor at each point in time using CS optimization. To assess performance, we consider decoding accuracy only during odor whiffs, rather than blanks, where the concentration is too low to be perceived. We assess performance by the percentage of correctly decoded whiffs (signal must be fully decoded at some point during the whiff), and average our results over distinct choices of foreground and background identity. Finally, we assume the decoder has shortterm memory: detected odor signals are only retained for ${\tau}_{\text{M}}$ seconds in the immediate past, bounding the amount of past information utilized in signal reconstruction.
Without ORN adaptation, sufficiently strong backgrounds eliminate the ability to reconstruct the identity of individual odor whiffs, irrespective of the complexity of either the foreground or background odor (Figure 3E, blue lines). In the adaptive system, this is substantially mitigated (red lines in Figure 3E), provided the memory duration ${\tau}_{\text{M}}$ is at least as long as the adaptation timescale $\tau $ (darker red lines). The memory ${\tau}_{\text{M}}$ must be long enough so that information about the background concentration ${\overline{s}}_{i}$, which is needed for decoding, can be acquired over a window at least as long as the adaptation timescale. Because shortterm adaptation depends on the activity of the OrOrco channel rather than on the identity of the receptor (Nagel and Wilson, 2011; Martelli et al., 2013; GorurShandilya et al., 2017), the values of $\tau $ and ${A}_{0a}$ were assumed the same for all ORNs; still, our results hold if these invariances are relaxed (Figure 3—figure supplement 1 and Figure 3—figure supplement 2).
Frontend adaptation enhances primacy coding
The primacy coding hypothesis has recently emerged as an intriguing framework for combinatorial odor coding. Here, odor identity is encoded by the set (but not temporal order) of the $p$ earliest responding glomeruli/ORN types, known as primacy set of order $p$(Wilson et al., 2017). If the activation order of ORNs were invariant to the strength of an odor step or pulse, primacy sets would in principle form concentrationinvariant representation of odor identity. Although our coding framework uses the full ORN ensemble in signal reconstruction, some of these responses may contain redundant information, and a smaller primacy subset may suffice. To examine this, we apply our model to a sigmoidal stimulus that rises to halfmax in 50 ms, calculating decoding accuracy in time. Since ORNs activate sequentially, the primacy set is defined by the ORN subset active when the odor is decoded. For simple odors, a limited set of earliest responding neurons fully accounts for the odor identity (Figure 4A), in agreement with primacy coding. As expected for more complex odor mixtures, the full repertoire is required for accurate decoding. Primacy coding also predicts that for stronger stimuli, responses occur earlier, since the primacy set is realized quicker, which our framework replicates (Figure 4—figure supplement 1).
Beyond mere consistency, however, frontend adaptation might also enhance primacy coding in different environments, such as background odors, which could scramble primacy sets. To investigate this, we considered again a sigmoidal odor step (odor A), now atop a static background (odor B) to which the system has adapted. We compared the primacy sets of odor A for 1000 different choices of odor B, finding that, with adaptation, primacy sets are highly consistent across background confounds for all but the smallest primacy orders (Figure 4BFigure 4C). This also holds true for backgrounds of different concentrations (Figure 4—figure supplement 1), suggesting a central role for frontend adaptation in reinforcing primacy codes across differing environmental conditions.
Contribution of frontend adaptation for odor recognition within the Drosophila olfactory circuit
Signal transformations in the sensing periphery are propagated through the remainder of the olfactory circuit. How does frontend adaptation interact with these subsequent neural transformations? ORNs expressing the same OR converge to a unique AL glomerulus, where they receive lateral inhibition from other glomeruli (Olsen and Wilson, 2008; Asahina et al., 2009). This inhibition implements a type of divisive gain control (Olsen et al., 2010), normalizing the activity of output projections neurons, which then synapse onto a large number of Kenyon cells (KCs) in the mushroom body. To investigate how odor representations are affected by interactions between frontend ORN adaptation and this lateral inhibition and synaptic divergence, we extended our ORN encoding model by adding uniglomerular connections from ORNs to the antennal lobe, followed by sparse, divergent connections to 2500 KCs (Keene and Waddell, 2007; LitwinKumar et al., 2017; Caron et al., 2013). Inhibition was modeled via divisive normalization, with parameters chosen according to experiment (Olsen et al., 2010). We quantified decoding accuracy by training and testing a linear classifier on the KC activity output of sparse odors of distinct intensity and identity. We trained the classifier on ${N}_{\text{ID}}$ sparse odor identities at intensities chosen randomly over 4 orders of magnitude, then tested the classifier accuracy on the same set of odor identities but of differing concentrations.
With both ORN adaptation and divisive normalization, the accuracy of the classification by odor identity remains above 85% for more than 1000 odor identities (${N}_{\text{ID}}>1000$). Removing ORN adaption while maintaining divisive normalization substantially reduces accuracy (down to 65% for 1000 odor identities). Further removing divisive normalization gives similar results, apart for very large numbers of odors identities (${N}_{\text{ID}}>1000$), where divisive normalization provides benefits (Figure 5A). These results strongly implicate frontend adaptation as a key player in maintaining odor identity representations, before signals are further processed downstream.
As a simpler task, we also considered binary classification, categorizing odors as appetitive or aversive. For simplicity, odor signals of the same identity but differing intensity were assigned the same valence. Classification accuracy degrades to chance level as ${N}_{\text{ID}}$ becomes very large (Figure 5B). When acting alone, either divisive normalization or ORN adaptation can help, although the effect of ORN adaptation is slightly stronger. When both are active, accuracy improves further, suggesting that these distinct adaptive transformations may act jointly at different stages of neural processing in preserving representations of odor identity. As expected, these gains mostly vanish for the same odors chosen from a narrower range of concentrations (Figure 5—figure supplement 1).
Previous simulation results have shown that divisive normalization aids identity decoding from PN response to a stronger degree than we find here (Olsen et al., 2010). There, 19 distinct odor identities at three concentrations were classified more accurately with divisive normalization (80%) than without (68%). In our case, we find about ∼75% accuracy, with and without divisive normalization. This discrepancy is not necessarily inconsistent. First, we decode mixtures, not single odorants, and the combinatorics may reduce the benefit of divisive normalization. Second, we classify the responses of 2500 KCs, rather than 50 PNs (or 24 PNs as in Olsen et al., 2010). Kenyon cell responses follow a high degree of postsynaptic divergence from PNs, which could decorrelate neural responses (Caron et al., 2013; LitwinKumar et al., 2017; Krishnamurthy et al., 2017) similarly to divisive normalization, reducing the gains from the latter. Finally, the divisive normalization model is a simple one in which glomeruli are all mutually inhibiting. A more complex model in which each glomerulus inhibits only a subset of other glomeruli through local neurons might produce a larger contribution.
In sum, these results indicate that ORN adaptation might contribute significantly to odor recognition by identity and valence, and that divisive normalization also contributes, although possibly more to classification by valence than by identity. An intriguing possibility is that these two forms of gain control play different roles in coding discrete odor categories versus odor identities.
Discussion
WeberLaw adaptation at the very frontend of the insect olfactory circuit (GorurShandilya et al., 2017; Cafaro, 2016; Cao et al., 2016) may contribute significantly to the preservation of neural representations of odor identity amid confounding odors and intensity fluctuations. Drawing on experimental evidence for a number of ORNinvariant response features (Nagel and Wilson, 2011; Martelli et al., 2013; Stevens, 2016; GorurShandilya et al., 2017; Si et al., 2019), we have found that this mechanism of dynamic adaptation confers significant benefits in coding fidelity, without the need for ORNspecific parameterizations. Still, our results hold when these invariances such as adaptation timescale or baseline activity are relaxed (Figure 3—figure supplement 1 and Figure 3—figure supplement 2). In the olfactory periphery, frontend Weber Law adaptation therefore appears fairly robust, a consequence of controlling gain via feedback from channel activity (Waite et al., 2018; Nagel and Wilson, 2011; GorurShandilya et al., 2017), rather than through intrinsic, receptordependent mechanisms.
Our results also suggest that a slight breaking of Weber scaling may aid combinatorial coding, by spreading firing rates more fully over the ORN dynamic range, while still preventing saturation. The degree of this breaking would manifest as a correction to the Weber scaling exponent, $\sim {(1/s)}^{1}\to \sim {(1/s)}^{1\beta}$, which could in principle be measured experimentally for individual ORNs. Such small deviations from the strict WeberFechner scaling have been observed (see extended figures in GorurShandilya et al., 2017).
While our framework incorporates many observed features of the Drosphila olfactory system – WeberLaw adaptation, powerlaw distributed receptor affinities, temporal filter invariance, connectivity topologies – it is minimal. We considered only one of the chemoreceptor families expressed in the fly antenna (Joseph and Carlson, 2015) and ignored possible contributions of odor binding proteins (Vogt and Riddiford, 1981; Menuz et al., 2014), inhibitory odorants (Cao et al., 2017), and odorantodorant antagonism (Reddy et al., 2018), which could further boost coding capacity and preserve representation sparsity. Useful extensions to our nonlinearlinearnonlinear model might incorporate ephaptic coupling between ORNs housed in the same sensillum (Su et al., 2012), global inhibition in the mushroom body (Papadopoulou et al., 2011), and the effects of longterm adaptation (Guo et al., 2017).
Previous studies have characterized various neural mechanisms that help preserve combinatorial codes. Lateral inhibition between glomeruli helps tame saturation and boost weak signals (Olsen et al., 2010). The sparse degree of connectivity to either the olfactory bulb (vertebrates) or mushroom body (insects) may also be precisely tuned to optimize the capacity to learn associations (LitwinKumar et al., 2017). In this work, we find that some of these downstream features act in concert with frontend dynamic adaptation in maintaining representations of odor identity.
Other studies have implicated the unique temporal patterns of neural response as signatures of odor identity (Raman et al., 2010; Gupta and Stopfer, 2011; Brown et al., 2005; Gupta and Stopfer, 2014). ORN and projection neuron time traces form distinct trajectories in lowdimensional projections, and cluster by odor identity, much as we have found here for static responses at different concentrations (Figure 2). In locusts PNs, the trajectories elicited by foreground odors when presented in distinct backgrounds exhibit some degree of overlap; although partial, these overlaps were nonetheless sufficient to maintain backgroundinvariant decoding from Kenyon cell responses (Saha et al., 2013). It was therefore suggested that background filtering likely occurs at the level of ORNs themselves (Saha et al., 2013). Likewise, in our framework, temporal coding is implicit: because the input nonlinearity depends on the diversity of binding affinities, odor signals are naturally formatted into temporal patterns that are both odor and ORNspecific (Figure 1DFigure 1E). Further, the short required memory timescales ($\tau}_{M}\sim \tau \sim 250\phantom{\rule{thinmathspace}{0ex}}\mathrm{m}\mathrm{s$) suggest that only brief time windows are needed for accurate odor identification, consistent with previous findings (Brown et al., 2005; Saha et al., 2013). Moreover, we find that frontend adaptation enhances the robustness of other combinatorial coding schemes, such as primacy coding (Wilson et al., 2017), which relies on the temporal order of ORN activation but not absolute firing rate (Figure 4).
In the wellcharacterized chemosensory system of bacterial chemotaxis, Weber Law adaptation is enacted through a feedback loop from the output activity of the receptorkinase complexes onto the enzymes modifying receptor sensitivity (Waite et al., 2018). It is interesting that some aspects of this logic are also present in ORNs: although the molecular players are different (and still largely unknown, although likely involving calcium channel signaling, Cao et al., 2016), it has been shown that transduction activity feeds back onto the sensitivity of OrOrco ligandgated cation channels, enabling the WeberFechner relation (Nagel and Wilson, 2011; GorurShandilya et al., 2017; Cao et al., 2016). That this adaptation mechanism appears to act similarly across ORNs (GorurShandilya et al., 2017; Martelli et al., 2013; Cao et al., 2016) suggests the possible involvement of the universal coreceptor Orco, whose role in longterm adaptation has recently been reported (Getahun et al., 2013; Getahun et al., 2016; Guo et al., 2017). Further, the identification of four subunits comprising the OrcoOr ion channel suggest that generic Or/Orco complexes may contain multiple odorant binding sites, which when included in our model supports our general findings (Figure 3—figure supplement 3).
Weber Law ensures that sensory systems remain in the regime of maximum sensitivity, broadening dynamic range and maintaining information capacity (Wark et al., 2007). For a singlechannel system, this requires matching the midpoint of the doseresponse curve to the mean ligand concentration (Nemenman, 2012), a strategy which may fail in multichannel systems with overlapping tuning curves: adaptation to one signal could inhibit identification of others, if the signals excite some but not all the same sensors, as in Figure 1G. Our results show that this strategy is still largely functional. In CS decoding, this can be traced to the observation that accuracy is guaranteed when sufficiently distinct odor identities produce sufficiently distinct ORN responses, a condition known as the restricted isometry property (Candès et al., 2006). Indeed, the WeberFechner scaling increases the likelihood that this property is satisfied, beyond that in the nonadaptive system (SI text and Figure 3—figure supplement 4  Figure 3—figure supplement 5). Still, restricted isometry does not require that response repertoires are invariant to environmental changes. That is, even if the subset of active ORNs were concentrationdependent, odors could still in principle be fully reconstructible by CS. Nonetheless, our results in tSNE clustering (Figure 2), primacy coding (Figure 4B–4C), and odor classification (Figure 5) suggest that some signature of response invariance emerges as a natural byproduct of frontend adaptation. Together, this implies that Weber Law adaptation, whether required by the olfactory circuit for precise signal reconstruction (as in CS) or for developing odor associations (as in classification), can play an integral part in maintaining combinatorial codes amid changing environmental conditions.
Materials and methods
Adaptive ORN model
Request a detailed protocolWe model an odor as an $N$dimensional vector $\mathbf{\mathbf{s}}=[{s}_{1},\mathrm{\dots},{s}_{N}]$, where ${s}_{i}>0$ are the concentrations of individual volatile molecules (odorants) comprising the odor. The olfactory sensory system is modeled as a collection of $M$ distinct Or/Orco complexes indexed by the sub index $a=1,\mathrm{\dots},M$, each of which can be bound with any one of the odorant molecules, and can be either active (firing) or inactive (quiescent). At first, we assume there is one binding site per complex; this will be generalized to many sites. We consider the binding and activation processes to be in equilibrium, assigning each state a corresponding Boltzmann weight, where the zero of energy is set by the unbound, inactive state ${\text{C}}_{a}$. These weights are:
where ${\u03f5}_{a}$ (assumed positive) is the free energy difference between the active and inactive conformation of the unbound receptor, and ${E}_{ai}$ and ${E}_{ai}^{*}$ are the free energy differences (assumed positive) between the unbound and bound state for the inactive and active receptor, respectively. ${\mu}_{i}={\mu}_{0}+{\beta}^{1}\mathrm{log}({s}_{i}/{s}_{0})$ is the chemical potential for odorant species $i$ in terms of a reference chemical potential ${\mu}_{0}$ at concentration ${s}_{0}$, ${s}_{0}\mathrm{exp}(\beta {\mu}_{0})={s}_{i}\mathrm{exp}(\beta {\mu}_{i})$, which can be traded for the thermodynamicrelevant disassociation constants ${K}_{ai}^{1}={K}_{\text{D},ai}={s}_{0}{e}^{\beta ({E}_{ai}{\mu}_{0})}$.
Adding up contributions from all $i$ odorants, the active fraction is:
where we have expressed free energies in units of ${k}_{B}T={\beta}^{1}$ for notational convenience.
This expression can be generalized for the case of multiple, independent binding sites through some simple combinatorial factors. Consider first an odorant $i$ which can bind one of two locations on receptor $a$. There are then four possible inactive states: both sites unbound, site one bound, site two bound, both sites bound. Combined with the active states, there are therefore eight states for odorant $i$ and receptor $a$, with energies:
In the active fraction, Equation 2, the Boltzmann factors combine through the binomial theorem, giving (for a single odorant environment $i$):
This expression generalizes for an arbitrary number of odorants and independent binding sites through the appropriate combinatorial factors, giving an active fraction of
To generate ORN time traces, Equations 23 are integrated numerically using the Euler method with a 2 ms time step. For ORN firing (Equation 4), $h(t)$ is bilobed (Martelli et al., 2013): $h(t)=A{p}_{\mathrm{G}\mathrm{a}\mathrm{m}}(t;{\alpha}_{1},{\tau}_{1})B{p}_{\mathrm{G}\mathrm{a}\mathrm{m}}(t;{\alpha}_{2},{\tau}_{2})$, $A=190$, $B=1.33$, ${\alpha}_{1}=2$, ${\alpha}_{2}=3$, ${\tau}_{1}=0.012$, and ${\tau}_{2}=0.016$, where $p}_{\mathrm{G}\mathrm{a}\mathrm{m}$ is the pdf of Gamma($\alpha $, $1/\tau $). Nonlinearity $f$ is modeled as a linear rectifier with 5 Hz threshold.
Derivation of ORN gain
Request a detailed protocolWeber’s Law states that the gain, or differential response, of the receptor activity ${A}_{a}$ scales with the mean odor concentration ${\overline{s}}_{i}$. To show how this is satisfied in our model, we consider the response, Equation 2, to a signal $\mathbf{\mathbf{s}}=\overline{\mathbf{\mathbf{s}}}+\mathrm{\Delta}\mathbf{\mathbf{s}}$, where $\mathrm{\Delta}\mathbf{\mathbf{s}}$ consists of only a small fluctuation in the $i$th component $\mathrm{\Delta}{s}_{i}<{\overline{s}}_{i}$ about the mean. We derive the change in response to fluctuation $\mathrm{\Delta}{s}_{i}$ for general $\beta $ from 0 (Weber’s law) to 1 (no adaptation).
First we write the activity in the form:
where
where ${\u03f5}_{a}(\overline{\mathbf{\mathbf{s}}})$ is given by Equation 5. Then, assuming $1/{\mathbf{\mathbf{K}}}_{a}^{*}\ll {s}_{i}\ll 1/{\mathbf{\mathbf{K}}}_{a}$, the change in response from the adapted level ${A}_{a}(\overline{\mathbf{\mathbf{s}}})$ is
We use Equation 5 to evaluate ${e}^{{F}_{a}}$ at $\overline{\mathbf{\mathbf{s}}}$, obtaining:
whereby
For $\beta =0$ (the fully adaptive case) and a single odorant, this expression for the gain reduces to $(1{A}_{0a}){A}_{0a}/{s}_{i}$. For small $\beta $, and given ${A}_{0a}\simeq 0.1$ (corresponding to 30 Hz on a 300 Hz firing rate scale), the denominator is dominated by the $1{A}_{0a}$ term, giving:
The implication of this is that the gain scaling of the inverse mean intensity, which is 1 for perfect adaptation (gain $\sim {(1/{s}_{i})}^{1}$), is now sublinear. Thus, when Weber’s Law is weakly broken, the gain still reduces with mean odor intensity, but not as quickly.
tSNE dimensionality reduction and mutual information
Request a detailed protocolFor tSNE dimensionality reduction (van der Maaten and Hinton, 2008), ORN responses were generated for odor signal combinations consisting of 1 (among 10) distinct sparse foreground odors A atop 1 (among 50) distinct sparse background odors B, for Figure 2B. Figure 2C plots responses for 10 odors at 40 concentrations spanning four decades, atop a random sparse background odor of similar magnitude. For adaptive systems, ${\u03f5}_{a}$ were set to their fully adapted values to the background odor, given by Equation 5, with $\beta =0$.
The mutual information (MI) between signal and response quantifies how many bits of information a response contains about the stimulus. As we are interested in how much information ORN responses $\mathbf{\mathbf{r}}$ contain about novel foreground odors $\mathbf{\mathbf{s}}$, we calculate the MI between $\mathbf{\mathbf{s}}$ and $\mathbf{\mathbf{r}}$. This calculation requires the conditional response distribution $P(\mathbf{\mathbf{r}}\mathbf{\mathbf{s}})$, where the probability distribution is over different background odors $\overline{\mathbf{\mathbf{s}}}$. To get this distribution, we hold $\mathbf{\mathbf{s}}$ fixed and generate $\mathbf{\mathbf{r}}$ in the presence of distinct backgrounds. To these responses $\mathbf{\mathbf{r}}$, we also add a small amount of Gaussian noise (mean zero and variance 1 Hz), which allows a distribution to be defined when there is no background odor. We then bin the resulting $\mathbf{\mathbf{r}}$ in units of $\mathrm{\Delta}r$ = 1 Hz to get a histogram representing $P(\mathbf{\mathbf{r}}\mathbf{\mathbf{s}})$ (The histogram is necessary, since a sum must be taken over $\mathbf{\mathbf{r}}$). If responses were completely background invariant, the resulting histogram would be highly peaked.
Using $P(\mathbf{\mathbf{r}}\mathbf{\mathbf{s}})$, the MI is defined as
where $H}_{\mathrm{n}\mathrm{o}\mathrm{i}\mathrm{s}\mathrm{e}$ is:
and $H}_{\mathrm{r}\mathrm{e}\mathrm{s}\mathrm{p}\mathrm{o}\mathrm{n}\mathrm{s}\mathrm{e}$ is
where
The noise entropy $H}_{\mathrm{n}\mathrm{o}\mathrm{i}\mathrm{s}\mathrm{e}$ quantifies how much variability comes from the background odors, but is not related to changes in foreground odor. The response entropy $H}_{\mathrm{r}\mathrm{e}\mathrm{s}\mathrm{p}\mathrm{o}\mathrm{n}\mathrm{s}\mathrm{e}$ quantifies how much variability comes from both background and foreground. The mutual information, which is their difference, is a measure of how responses differ by foreground alone.
Compressed sensing decoding of ORN responses
Request a detailed protocolCompressed sensing (CS) addresses the problem of determining a sparse signal from a set of linear measurements, when the number of measurements is less than the signal dimension. Specifically, it is a solution to
where $\mathbf{\mathbf{x}}\in {\mathbb{R}}^{N}$ and $\mathbf{\mathbf{y}}\in {\mathbb{R}}^{M}$ are vectors of signals and responses, respectively, and $\mathbf{\mathbf{D}}$ is the measurement matrix. Since measurements are fewer than signal components, then $M<N$, whereby $\mathbf{\mathbf{D}}$ is wide rectangular and so Equation 16 cannot be simply inverted to produce $\mathbf{\mathbf{x}}$. The idea of CS is to utilize the knowledge that $\mathbf{\mathbf{x}}$ is sparse, that is only $K$ of its components, $K\ll N$ are nonzero. Both the measurements and sparsity are thus combined into a single constrained optimization routine:
where ${\widehat{x}}_{i}$ are the optimal estimates of the signal components and the sum, which is known as the ${L}_{1}$ norm of $\mathbf{\mathbf{x}}$, is a natural metric of sparsity (Donoho, 2006).
The ${L}_{1}$ norm is a convex operation and the constraints are linear, so the optimization has a unique global minimum. To incorporate the nonlinear response of our encoding model into this linear framework, we assume that the responses are generated through the full nonlinear steady state response, Equations 2 4, but that the measurement matrix $\mathbf{\mathbf{D}}$ needed for decoding uses a linear approximation of this transformation. Expanding Equation 4 around $\overline{\mathbf{\mathbf{s}}}=\mathbf{\mathbf{s}}\mathrm{\Delta}\mathbf{\mathbf{s}}$ gives
where
and where ${\frac{d{A}_{ai}}{d{s}_{i}}}_{\overline{\mathbf{\mathbf{s}}}}$ is given by the righthand side of Equation 14 with $\beta =0$. Equations 18 and 19 hold only for integrands above 5 Hz (and are zero below), as per the linear rectifier $f$. We assume that the neural decoder has access to background $\overline{\mathbf{\mathbf{s}}}$, presumed learned (this assumption can be relaxed; see below), and to the linearized response matrix, Equation 14, but must infer the excess signals $\mathrm{\Delta}{s}_{i}$ from excess ORN firing rates $\mathrm{\Delta}{r}_{a}(t)$. Thus, this corresponds to the CS framework (Equation 17) via $\mathrm{\Delta}\mathbf{\mathbf{r}}\to \mathbf{\mathbf{y}}$, $\mathrm{\Delta}\mathbf{\mathbf{s}}\to \mathbf{\mathbf{x}}$, and ${d{A}_{ai}/d{s}_{i}}_{\overline{\mathbf{\mathbf{s}}}}\to \mathbf{\mathbf{D}}$. We optimize the cost function in Equation 17 using sequential least squares programming, implemented in Python through using the scientific package SciPy.
For our simulations, we let sparse components ${s}_{i}$ be chosen as ${s}_{i}={\overline{s}}_{i}+\mathrm{\Delta}{s}_{i}$, where ${\overline{s}}_{i}={s}_{0}$ and $\mathrm{\Delta}{s}_{i}\sim \mathcal{N}({s}_{0}/3,{s}_{0}/9)$. The measurement matrix $\mathbf{\mathbf{D}}$ depends on the free energy differences ${\u03f5}_{a}$. For static stimuli, ${\u03f5}_{a}$ equals the fixed point of Equation 3 in response to the background stimulus with $\beta =0$. For fluctuating stimuli, ${\u03f5}_{a}$ is updated in time by continuously integrating ${r}_{a}(t)$, via Equation 3 and 4; thus, only knowledge of the response ${r}_{a}(t)$ are needed by the decoder. To quantify decoding accuracy, we treat the zero and nonzero components of the sparse odor vector separately. We demand that the $K$ nonzero components ${\widehat{s}}_{i}$ of the estimated sparse vector are within 25% of their true values ${s}_{i}$, and that the $NK$ zero components are estimated less than 10% of ${s}_{0}$. Together, this ensures that the odorants comprising the odor mixture are estimated sufficiently close to their concentrations, and that the remaining components are sufficiently small. Odor signals $\mathbf{\mathbf{s}}$ are considered correctly decoded if both of these conditions are satisfied for all components ${s}_{i}$. The relatively lax accuracy demanded on the nonzero components is to prevent oversensitivity on the unavoidable errors introduced by linearization. Qualitatively, our findings are robust to these choices.
The naturalistic odor signal (Figure 3D) was generated by randomly varying flow rates of ethyl acetate and measuring the concentration with a photoionization detector (GorurShandilya et al., 2017). Statistics mirroring a turbulent flow (Celani et al., 2014) were verified (Figure 3—figure supplement 6).
Iterative hard thresholding (IHT) and the restricted isometry property in compressed sensing
Request a detailed protocolThe purpose of response linearization (Equation 18) is simply to apply compressed sensing reconstruction directly using linear programming, without worrying about issues of local minima in Equation 17. This allows us to isolate the impact of Weber Law adaptation from the particularities of the numerics. An alternate technique for compressed signal reconstruction, iterative hard thresholding (IHT), does not minimize the constrained ${L}_{1}$ norm directly, rather applying a hard threshold to an iteratively updated signal estimate (Blumensath and Davies, 2009b). IHT can be generalized straightforwardly to nonlinear constraints, and would actually dispense with the need for a learned background $\overline{\mathbf{\mathbf{s}}}$, simply initializing the iterations from $\overline{\mathbf{\mathbf{s}}}=\mathrm{\U0001d7ce}$. Remarkably, this technique works quite well even for nonlinear measurements (Blumensath, 2013). We demonstrate the applicability of the IHT algorithm to our odor decoding system in Figure 3—figure supplement 5, which reproduces qualitatively the findings in the main text. For these calculations, no background odor was assumed, each iterative decoding being initialized $\overline{\mathbf{\mathbf{s}}}=\mathrm{\U0001d7ce}$.
IHT provides an alternate computational technique of nonlinear CS, which could be used to both extend and verify our results. Further, it allows us to illustrate why Weber Law adaptation maintains signal reconstruction fidelity in our olfactory sensing model. Like CS using ${L}_{1}$norm minimization, IHT exhibits amenable reconstruction and convergence properties under the guarantee of the socalled restricted isometry property (RIP) (Candès et al., 2006). Loosely, RIP measures how closely a matrix operator resembles an orthogonal transformation when acting on sparse vectors. The degree to which RIP is satisfied can be understood in terms of the spectrum of a measurement matrix $\mathbf{\mathbf{D}}$. In particular, if ${\lambda}_{i}$ are the eigenvalues of ${\mathbf{\mathbf{D}}}_{i}^{T}{\mathbf{\mathbf{D}}}_{i}$, where ${\mathbf{\mathbf{D}}}_{i}$ is any $k\times m$ submatrix of $\mathbf{\mathbf{D}}$, and
is satisfied for some ${\delta}_{i}$, then $\mathbf{\mathbf{D}}$ satisfies the RIP with constant ${\delta}_{i}$. Plainly, the RIP states that the eigenvalues of ${\mathbf{\mathbf{D}}}_{i}^{T}{\mathbf{\mathbf{D}}}_{i}$, when acting on $k$sparse vectors, are centered around 1. Thus, to intuit why signal reconstruction breaks down in the nonadaptive sensing system, we can investigate the eigendecomposition of various linearizations of the measurement matrix. We do this now, starting with a brief description of the IHT.
In the linear setting, IHT seeks sparse signals via the following iterative procedure (Blumensath and Davies, 2009b):
where ${\mathbf{\mathbf{s}}}_{i}$ is the $i$th estimate of the sparse signal $\mathbf{\mathbf{s}}$, $\mu $ is a step size for the iterations, and $\mathbf{\mathbf{y}}$, $\mathbf{\mathbf{D}}$ are as defined above. ${H}_{k}(\cdot )$ is a thresholding function which sets all but the largest $K$ values of its argument to zero. The nonlinear extension to IHT is (Blumensath, 2013):
where $D$ is a nonlinear sensing function and ${\mathbf{\mathbf{D}}}_{{\mathbf{\mathbf{s}}}_{i}}$ is a linearization of $D$ about the point ${\mathbf{\mathbf{s}}}_{i}$. Reconstructibility for $k$sparse signals is guaranteed if ${\mathbf{\mathbf{D}}}_{{\mathbf{\mathbf{s}}}_{i}}$ satisfies RIP for all ${\mathbf{\mathbf{s}}}_{i}$ and all $k$sparse vectors (Blumensath and Davies, 2009b). To get a sense of how this is preserved in the adaptive system, we calculate the eigenvalues for 1000 choices of ${\mathbf{\mathbf{s}}}_{i}$, acting on random signals of given sparsity $K$ (Figure 3—figure supplement 4). Since the RIP is sensitive to constant scalings of the measurement matrix (while the actual estimation problem is not), we scaled all columns of ${\mathbf{\mathbf{D}}}_{{\mathbf{\mathbf{s}}}_{i}}$ to norm unity (Blumensath and Davies, 2009a). This normalizes the eigenvalues of ${\mathbf{\mathbf{D}}}_{{\mathbf{\mathbf{s}}}_{i}}^{T}{\mathbf{\mathbf{D}}}_{{\mathbf{\mathbf{s}}}_{i}}$ to center near unity before calculating the eigendecomposition, allowing us to assess the degree to which the RIP is satisfied. This scaled matrix can be used directly in Equation 22 (Blumensath, 2013; Blumensath and Davies, 2009a). The spectra of these matrices indicates that the RIP becomes far more weakly satisifed in the nonadaptive system than in the adaptive one, for sufficient odor complexity and intensity.
Network model and classification
Request a detailed protocolFor the network model, the ALtoMB connectivity matrix ${\mathbf{\mathbf{J}}}_{1}$, is chosen such that each KC connects presynaptically to seven randomly chosen AL glomeruli (LitwinKumar et al., 2017; Caron et al., 2013). The results shown in Figure 5 are an average of 10 distinct instantiations of this random topology. The $Z=2500$ KCs are then connected by a matrix ${\mathbf{\mathbf{J}}}_{2}$ to a readout layer of dimension $Q$, where $Q=2$ for binary and $Q={N}_{\text{ID}}$ for multiclass classification. Both ALtoMB and MBtoreadout connections are perceptrontype with rectifiedlinear thresholds. The weights of ${\mathbf{\mathbf{J}}}_{1}$ and ${\mathbf{\mathbf{J}}}_{2}$ are chosen randomly from $\sim \mathcal{N}(0,1/\sqrt{7})$ and $\sim \mathcal{N}(0,1/\sqrt{Z})$, respectively. Only the ${\mathbf{\mathbf{J}}}_{2}$ and the MBtooutput thresholds are updated during supervised network training, via logistic regression (for binary classification) or its higherdimensional generalization, the softmax cross entropy (for multiclass classification).
References
 1

2
Compressed sensing with nonlinear observations and related nonlinear optimization problemsIEEE Transactions on Information Theory 59:3466–3474.https://doi.org/10.1109/TIT.2013.2245716

3
Proceedings of SPARS’09  Signal Processing with Adaptive Sparse Structured RepresentationsHow to use the iterative hard thresholding algorithm, Proceedings of SPARS’09  Signal Processing with Adaptive Sparse Structured Representations, France, Saint Malo.

4
Iterative hard thresholding for compressed sensingApplied and Computational Harmonic Analysis 27:265–274.https://doi.org/10.1016/j.acha.2009.04.002

5
Encoding a temporally structured stimulus with a temporally structured neural representationNature Neuroscience 8:1568–1576.https://doi.org/10.1038/nn1559
 6
 7
 8

9
Stable signal recovery from incomplete and inaccurate measurementsCommunications on Pure and Applied Mathematics 59:1207–1223.https://doi.org/10.1002/cpa.20124
 10
 11

12
Navigational strategies used by insects to find distant, windborne sources of odorJournal of Chemical Ecology 34:854–866.https://doi.org/10.1007/s1088600894845
 13

14
Odor landscapes in turbulent environmentsPhysical Review X 4:041015.https://doi.org/10.1103/PhysRevX.4.041015
 15
 16

17
A model of Drosophila larva chemotaxisPLOS Computational Biology 11:e1004606–e1004624.https://doi.org/10.1371/journal.pcbi.1004606
 18

19
Compressed sensingIEEE Transactions on Information Theory 52:1289–1306.https://doi.org/10.1109/TIT.2006.871582
 20
 21
 22

23
Intracellular regulation of the insect chemoreceptor complex impacts odour localization in flying insectsThe Journal of Experimental Biology 219:3428–3438.https://doi.org/10.1242/jeb.143396
 24

25
Odorant receptor sensitivity modulation in DrosophilaThe Journal of Neuroscience 37:9465–9473.https://doi.org/10.1523/JNEUROSCI.157317.2017

26
Odorant receptor desensitization in insectsJournal of Experimental Neuroscience 11:1–5.https://doi.org/10.1177/1179069517748600

27
Insect olfactory coding and memory at multiple timescalesCurrent Opinion in Neurobiology 21:768–773.https://doi.org/10.1016/j.conb.2011.05.005

28
A temporal channel for information in sparse sensory codingCurrent biology : CB 24:2247–2256.https://doi.org/10.1016/j.cub.2014.08.021
 29

30
Mechanisms of olfactory discrimination: converging evidence for common principles across phylaAnnual Review of Neuroscience 20:595–631.https://doi.org/10.1146/annurev.neuro.20.1.595
 31
 32

33
Drosophila olfactory memory: single genes to complex neural circuitsNature Reviews Neuroscience 8:341–354.https://doi.org/10.1038/nrn2098
 34
 35
 36
 37

38
Intensity invariant dynamics and odorspecific latencies in olfactory receptor neuron responseJournal of Neuroscience 33:6285–6297.https://doi.org/10.1523/JNEUROSCI.042612.2013
 39

40
Similar odorants elicit different behavioral and physiological responses, some supersustainedJournal of Neuroscience 31:7891–7899.https://doi.org/10.1523/JNEUROSCI.625410.2011

41
Odor plumes and how insects use themAnnual Review of Entomology 37:505–532.https://doi.org/10.1146/annurev.en.37.010192.002445

42
Biophysical mechanisms underlying olfactory receptor neuron dynamicsNature Neuroscience 14:208–216.https://doi.org/10.1038/nn.2725

43
A largescale analysis of odor coding in the olfactory epitheliumJournal of Neuroscience 31:9179–9191.https://doi.org/10.1523/JNEUROSCI.128211.2011

44
Information theory and adaptationIn: M. E Wall, editors. Quantitative Biology: From Molecular to Cellular Systems, 4. USA: CRC Press. pp. 73–91.
 45
 46
 47

48
A clustering neural network model of insect olfactionIn 2017 51st Asilomar Conference on Signals, Systems, and Computers. pp. 593–600.
 49
 50
 51

52
A spatiotemporal coding mechanism for backgroundinvariant odor recognitionNature Neuroscience 16:1830–1839.https://doi.org/10.1038/nn.3570
 53
 54
 55
 56
 57

58
Visualizing highdimensional data using tsneJournal of Machine Learning Research 9:2579–2605.
 59
 60
 61

62
Behavioral variability and phenotypic diversity in bacterial chemotaxisAnnual Review of Biophysics 47:595–616.https://doi.org/10.1146/annurevbiophys062215010954
 63

64
Sensory adaptationCurrent Opinion in Neurobiology 17:423–429.https://doi.org/10.1016/j.conb.2007.07.001
 65

66
The fluid dynamical context of chemosensory behaviorThe Biological Bulletin 198:188–202.https://doi.org/10.2307/1542523

67
Early olfactory processing in Drosophila: mechanisms and principlesAnnual Review of Neuroscience 36:217–241.https://doi.org/10.1146/annurevneuro062111150533

68
A primacy code for odor identityNature Communications 8:1477.https://doi.org/10.1038/s41467017014324
 69
Decision letter

Fred RiekeReviewing Editor; University of Washington, United States

Catherine DulacSenior Editor; Harvard University, United States

Katherine I. NagelReviewer; NYU, United States
In the interests of transparency, eLife includes the editorial decision letter and accompanying author responses. A lightly edited version of the letter sent to the authors after peer review is shown, indicating the most substantive concerns; minor comments are not usually included.
Thank you for submitting your article "Frontend WeberFechner gain control enhances the fidelity of combinatorial odor coding" for consideration by eLife. Your article has been reviewed by three peer reviewers, and the evaluation has been overseen by a Reviewing Editor and Catherine Dulac as the Senior Editor. The following individual involved in review of your submission has agreed to reveal their identity: Katherine I. Nagel (Reviewer #2).
The reviewers have discussed the reviews with one another and the Reviewing Editor has drafted this decision to help you prepare a revised submission.
The reviewers agreed that the work was of potential interest. They also agreed, however, that the presentation needs to be improved considerably. To be considered further, the paper needs to be revised carefully considering eLife's broad readership. This includes unpacking details about technical aspects of the work – such as embedding and compressed sensing – and explaining the key mathematical concepts of the model in words and (better) in schematic figures. The terseness with which the results were presented made evaluating the work difficult. The individual reviews below contain specific suggestions and comments that should help identify where such changes are needed, but we would also encourage the authors to ask for comments from several nonolfactory colleagues, particularly about accessibility.
Individual reviews follow:
Reviewer #1:
This paper investigates the impact of Weber adaptation in olfactory receptor neurons on olfactory coding using a model based on past experimental work (described in the paper this one is linked too). The central question should be of general interest, and the approach taken in the paper seems appropriate. I struggled, however, with the way the work is presented and this left me unsure about the conclusions reached. I am not an expert in olfaction, but I suspect these struggles will be shared by many other potential readers.
Response dynamics: I was quite confused about the importance of differences in response dynamics of different ORNs. In places the text appears to state that differences in dynamics are small (e.g. Introduction), and in others that they are important (subsection “Model of ORN sensing repertoire”). Some of this may originate from responses of a single cell to multiple odors vs. responses of different cells. Nonetheless, the present version of the paper is confusing in this regard.
Figure 2 and embedding: The embedding process used in the analysis illustrated in Figure 2 is not explained in any detail – meaning that I could not interpret Figure 2. Later in the Discussion (third paragraph) this figure is referred to with respect to response dynamics – this was particularly unclear. This figure is critical to the paper, so must be explained in more detail.
Figure 3: The use of compressed sensing in the decoding analysis in this figure is unclear. Related to this point, it's not clear how an appropriate tolerance is chosen (subsection “Frontend adaptation enhances odor discrimination in complex environments”, second paragraph). The approach to decoding needs to be described in considerable more detail.
Discrimination in complex odor environments: it is not clear here why the background should be represented as static. I would have thought it would be subject to many of the same properties that make the signal dynamic. The role/importance of short term memory is also unclear.
It would be interesting to see how important ORNspecific adaptation is for the results presented, as compared to a mechanism that acted universally across all ORN responses.
Equation 1: The origin of this equation could get explained in more detail.
Equation 2: This form of feedback, and particularly its relation to Weber adaptation, should get explained more.
Reviewer #2:
This manuscript asks how adaptation in olfactory receptor neurons (ORNs) impacts the ability of an olfactory system to encode odor identities reliably. There is a broad consensus in the field that odors are encoded by the combinatorial activity of an array of receptors, each composed of an odorspecific receptor and a common coreceptor. At least one form of adaptation, in which the sensitivity of olfactory receptor neurons is adjusted based on the activation level of the receptor complex, is present within ORNs, likely acting at the level of feedback onto the orco coreceptor. This study uses theoretical approaches to ask how this form of adaptation impacts decoding of odor identity, using three different models of odor decoding: compressed sensing, primacy coding, and a biologicallyinspired Kenyon cell model. The manuscript builds on a previous paper from the same group that developed a formulation for ORN adaptation based on a 2state receptor model. The broad finding of the study is that frontend adaptation improves odor identity decoding using a variety of models. Overall I think this study addresses an important question and does so in a thorough way, making use of very reasonable models for both odor encoding and decoding, and providing a nice overview of the state of the field. However, I think some elements of the exposition could be made more accessible for less mathematicallyinclined readers, and that some additional simulations would help pinpoint the reason why frontend adaptation improves encoding.
1) The manuscript is written for a highly quantitative audience and assumes a background familiar with the various models (receptor model, compressed sensing, tSNE) they employ. I think the paper could be made more accessible by unpacking some of the mathematical formulae in the main text.
For example, it would be helpful to show a plot of the activation function A_{a} as a function of odor concentration (Equation 1) for some of their sample model neurons, in both the unadapted and adapted state.
In addition, the discussion of compressed sensing is highly compressed. If the authors could describe this in an intuitive or graphical way in the main Results it would help readers understand what this is and how it works.
Using a KCinspired model to decode odor identity will probably be the most intuitive decoding scheme for many biologists. Here this decoding scheme is presented last but perhaps it might go earlier in the manuscript.
2) One possible interpretation of the results in Figures 2 and 3 is that in the nonadaptive system, high background odor concentrations cause the receptors to saturate, preventing them from encoding anything about the target odor, or at least massively compressing their dynamic range. This would mean that sensitivity adaptation is important (the activation curve needs to shift with increasing odor concentration), but not the precise form of the adaptation. Could the authors perform additional simulations to address this? For example: (1) What is the state of the receptors (distribution of activation levels) in the adapted versus unadapted system in high background odor (prior to target odor presentation) vs. background+target? (2) How do the results in Figures 2 and 3 differ if the adaptation is not exact? That is, what if there is some factor 𝛽 in front of A_{a}(t) in Equation 2? How precise does the adaptation have to be for this to work?
Reviewer #3:
The authors describe a receptor typeindependent adaptation mechanism at the level of the olfactory sensory neurons (OSNs) that maintains odor capacity in natural conditions. They proposed that adaptation or gain control follows the WeberFechner Law of psychophysics (previously shown by the same group) and suggest that in a biological context it may be driven by Orco coreceptor activity in a nonreceptor specific manner. The model results show that this kind of adaptation can aid concentrationinvariant coding, discrimination (even in the presence of background odors) as well as it agrees with the novel hypothesis of primacy coding. The topic discussed in the article is relevant and the results are convincing, it is worth publishing; I have no major concerns.
[Editors' note: further revisions were requested prior to acceptance, as described below.]
Thank you for submitting your article "Frontend WeberFechner gain control enhances the fidelity of combinatorial odor coding" for consideration by eLife. Your article has been reviewed by one peer reviewer, Fred Rieke, who is the Reviewing Editor and Reviewer #1, and the evaluation has been overseen by Catherine Dulac as the Senior Editor.
Reviewer #1:
This is a revision of a paper describing a modeling approach to explore the role of frontend adaptation in olfactory coding. The paper is interesting, and contains a number of nice analyses that provide insight into the interaction between adaptation and coding. The revisions have made the paper easier to understand, but there are still several issues that are not as clear as they need to be. These, and some smaller points, are detailed below. In general (as detailed below), for each analysis I think it is essential that each of the steps involved in going from the modeled responses to a completed piece of analysis need to be clear to a nonexpert reader.
1) tSNE analysis.
The comparison of tSNE to PCA (first paragraph of the subsection “Frontend WeberFechner adaptation preserves odor coding among background and intensity confounds”) is helpful. Can you build on the end of this paragraph to explain how tSNE works, and, critically, to define the axes of Figures 2B, C? It is quite important that a reader is comfortable with what is being plotted here.
2) CS analysis.
The description of the sparseness constraint added to the paper is helpful. What is still not clear, however, is how the stimulus itself is estimated (I can guess, but it should be stated explicitly). Related to this point, the signal perturbation (Ds) is defined only in the Materials and methods but needed to interpret the main text (subsection “Frontend adaptation enhances odor decoding in complex environments”, third paragraph). It should also be clearer that you are decoding discrete odor identify, not concentration (assuming that is correct).
3) Decoding timevarying stimuli.
Several aspects of the analysis described in the fifth paragraph of the subsection “Frontend adaptation enhances odor decoding in complex environments” are not clear. Were odors randomly assigned to each whiff? And was the entire time course decoded, or was each whiff treated as a discrete event? In general, the description of this analysis needs to be considerably more detailed. Is there an intuitive argument as to why the longer time scales of adaptation are helpful that could be added to the last paragraph of the subsection “Frontend adaptation enhances odor decoding in complex environments”?
4) Tests of primacy coding.
The text suggests that the background odors may interfere with primacy coding in the absence of frontend adaptation. The analysis presented in Figure 4 then shows that primacy sets are maintained in the presence of frontend adaptation. But there is not a test, unless I missed it, of the initial suggestion that primacy sets are not maintained without frontend adaptation. This test is needed to interpret this section.
5) Interplay of frontend adaptation and divisive normalization.
Figure 5 suggests that these two forms of gain control may play quite different roles in coding discrete odor categories (aversive, appetitive) and odor identity. This is quite interesting. I would consider swapping the order of presentation so you start with a discussion of odor identify (that flows more naturally from the previous sections). The differences between the present results and those of Olsen et al. should also, at a minimum, get discussed in more detail.
https://doi.org/10.7554/eLife.45293.020Author response
Reviewer #1:
This paper investigates the impact of Weber adaptation in olfactory receptor neurons on olfactory coding using a model based on past experimental work (described in the paper this one is linked too). The central question should be of general interest, and the approach taken in the paper seems appropriate. I struggled, however, with the way the work is presented and this left me unsure about the conclusions reached. I am not an expert in olfaction, but I suspect these struggles will be shared by many other potential readers.
Response dynamics: I was quite confused about the importance of differences in response dynamics of different ORNs. In places the text appears to state that differences in dynamics are small (e.g. Introduction), and in others that they are important (subsection “Model of ORN sensing repertoire”). Some of this may originate from responses of a single cell to multiple odors vs. responses of different cells. Nonetheless, the present version of the paper is confusing in this regard.
Our use of the wording “response dynamics” was confusing because it did not distinguish between two key contributions to ORN response: 1) odorreceptor binding and activation of the OROrco complex (Equation 2); 2) signal transduction and adaptation (Equations 34). Because odor binding and activation is nonlinear, variability in Step 1 introduces variability in the dynamic response of the ORN, even though the filter used for the firing rate is assumed the same for all ORNs. We edited the text to make clear that it is the signal transduction and adaptation dynamics that exhibit a surprising degree of invariance with respect to odorreceptor identity, not the odor binding and ion channel activation (subsection “Model of ORN sensing repertoire”, sixth paragraph).
Figure 2 and embedding: The embedding process used in the analysis illustrated in Figure 2 is not explained in any detail – meaning that I could not interpret Figure 2. Later in the Discussion (third paragraph) this figure is referred to with respect to response dynamics – this was particularly unclear. This figure is critical to the paper, so must be explained in more detail.
We rewrote portion of the text to better explain how the embedding is enacted, and why we use tSNE (versus PCA) to quantify the capability of the ORN repertoire to encode diverse odorants (subsection “Frontend WeberFechner adaptation preserves odor coding among background and intensity confounds”, first paragraph). We added a new panel A to Figure 2 to better introduce our approach and to help the reader interpret the other panels in the figure.
The later discussion about response dynamics is intended to draw parallels between our clustering results in Figure 2 and previous published results in which time traces of spiking activity were projected to a 3dimensional space. In both cases, responses cluster by odor identity. In these studies the authors used the entire time trace, while here we consider the response at a single time. We amended the text in the Discussion to clarify this (fifth paragraph).
Figure 3: The use of compressed sensing in the decoding analysis in this figure is unclear. Related to this point, it's not clear how an appropriate tolerance is chosen (subsection “Frontend adaptation enhances odor discrimination in complex environments”, second paragraph). The approach to decoding needs to be described in considerable more detail.
Indeed, our discussion of compressed sensing (CS) was too terse. We changed the first panel of Figure 3 to make it more intuitive, removing the unnecessary equations and replacing them with a simple graphic. We also added text to describe more fully the general compressed sensing framework (subsection “Frontend adaptation enhances odor decoding in complex environments”, first two paragraphs).
We have explained the details for the decoding tolerance precisely in the Materials and methods, explaining our choice for this tolerance, and noting that our results are robust to particular choices in the tolerance.
Discrimination in complex odor environments: it is not clear here why the background should be represented as static. I would have thought it would be subject to many of the same properties that make the signal dynamic. The role/importance of short term memory is also unclear.
We are concerned with the detection of novel odors amid odors already present, thereby assuming that backgrounds odors have persisted for some time beforehand. Given that the adaptation time for the adaptation mechanisms we discuss is on the order of 250ms, the background of odor needs not to be strictly static. If it evolved on a slower time scale it would be enough. We chose to simplify the presentation so that one of these odors is on a much slower timescale, effectively static. This may be conceivable if the foreground and background arise from spatially separated sources: e.g. a lawn may release a background “grass odor” everywhere, while a flower in that lawn releases a foreground “flower odor” localized in plumes streaming from the flower. Of course, there are other cases where one odor of interest fluctuates on the same timescale as another nuisance odor. Then the distinction between foreground and background is lost. In our framework these would be considered both foreground odors.
The role of short term memory is to limit the amount of information utilized from the past. We now mention this in the fifth paragraph of the subsection “Frontend adaptation enhances odor decoding in complex environments”.
It would be interesting to see how important ORNspecific adaptation is for the results presented, as compared to a mechanism that acted universally across all ORN responses.
Indeed. Thank you for the suggestion. Please see response to reviewer 2 comment #2.
Equation 1: The origin of this equation could get explained in more detail.
Equation 2: This form of feedback, and particularly its relation to Weber adaptation, should get explained more.
We have rewritten the text describing the model to provide more explanation and have added a step in the derivation of the former Equation (1) (now Equation 2) to make the derivation clearer. We have added paragraphs explaining the origin of Weber’s Law from the model, and two panels to Figure 1 to further illustrate the properties of the model and the Weber Law adaptation (subsection “Model of ORN sensing repertoire”;Figure 1F1G).
Reviewer #2:
[…] 1) The manuscript is written for a highly quantitative audience and assumes a background familiar with the various models (receptor model, compressed sensing, tSNE) they employ. I think the paper could be made more accessible by unpacking some of the mathematical formulae in the main text.
For example, it would be helpful to show a plot of the activation function A_{a} as a function of odor concentration (Equation 1) for some of their sample model neurons, in both the unadapted and adapted state.
In addition, the discussion of compressed sensing is highly compressed. If the authors could describe this in an intuitive or graphical way in the main Results it would help readers understand what this is and how it works.
We have rewritten large parts of the paper to make this clearer. Please see our second, third and last responses to reviewer 1’s questions.
Using a KCinspired model to decode odor identity will probably be the most intuitive decoding scheme for many biologists. Here this decoding scheme is presented last but perhaps it might go earlier in the manuscript.
We were also somewhat on the fence in the ordering of the results. We opted for this presentation mainly because primacy coding and compressed sensing decoding are more easily interpretable and far more tractable computationally without the added machinery of the AL and MB connectivity. Further, primacy coding has been shown in projection neurons, one step away from ORNs, so we presented it before we discuss the ALMB connectivity. We do note in the CS section that we will later investigate the implications of circuit mechanisms in later sections. For these reasons, we chose to keep the ordering as is.
2) One possible interpretation of the results in Figures 2 and 3 is that in the nonadaptive system, high background odor concentrations cause the receptors to saturate, preventing them from encoding anything about the target odor, or at least massively compressing their dynamic range. This would mean that sensitivity adaptation is important (the activation curve needs to shift with increasing odor concentration), but not the precise form of the adaptation. Could the authors perform additional simulations to address this? For example: (1) What is the state of the receptors (distribution of activation levels) in the adapted versus unadapted system in high background odor (prior to target odor presentation) vs. background+target? (2) How do the results in figures 2 and 3 differ if the adaptation is not exact? That is, what if there is some factor 𝛽 in front of A_{a}(t) in Equation 2? How precise does the adaptation have to be for this to work?
Thank you for this suggestion. This comment (and reviewer #1 last comment) suggests a need to investigate how much we can break Weber scaling and still maintain combinatorial codes. We have now extended the section on odor coding to address this issue (subsection “Frontend WeberFechner adaptation preserves odor coding among background and intensity confounds”). We have introduced in our model a new parameter 𝛽 that allows us to gradually break the WeberFechner’s scaling. When 𝛽=0 Weber’s law is strictly satisfied and when 𝛽=1 there is no adaptation. Increasing 𝛽 away from zero introduces a dependency of the adapted state on the background odor concentration. We added two panels in Figure 2 illustrating this.
[Editors' note: further revisions were requested prior to acceptance, as described below.]
Reviewer #1:
This is a revision of a paper describing a modeling approach to explore the role of frontend adaptation in olfactory coding. The paper is interesting, and contains a number of nice analyses that provide insight into the interaction between adaptation and coding. The revisions have made the paper easier to understand, but there are still several issues that are not as clear as they need to be. These, and some smaller points, are detailed below. In general (as detailed below), for each analysis I think it is essential that each of the steps involved in going from the modeled responses to a completed piece of analysis need to be clear to a nonexpert reader.
1) tSNE analysis.
The comparison of tSNE to PCA (first paragraph of the subsection “Frontend WeberFechner adaptation preserves odor coding among background and intensity confounds”) is helpful. Can you build on the end of this paragraph to explain how tSNE works, and, critically, to define the axes of Figures 2B, C? It is quite important that a reader is comfortable with what is being plotted here.
We have expanded our description of tSNE as suggested. tSNE only preserves local distances but not global distances. Thus, while it is a useful tool to cluster objects by similarity, the distances between clusters in the tSNE projection are not necessarily meaningful because global distances are not preserved. We now mention this both in the main text and in the caption. We also explain in the main text that we use tSNE primarily as a visualization method. To more rigorously quantify how well representations of odor identity are preserved, we calculate the mutual information (MI) between novel foreground odors and ORN responses in the 50dimensional space. We also provide a short explanation of what MI is in the main text.
2) CS analysis.
The description of the sparseness constraint added to the paper is helpful. What is still not clear, however, is how the stimulus itself is estimated (I can guess, but it should be stated explicitly). Related to this point, the signal perturbation (Ds) is defined only in the Materials and methods but needed to interpret the main text (subsection “Frontend adaptation enhances odor decoding in complex environments”, third paragraph). It should also be clearer that you are decoding discrete odor identify, not concentration (assuming that is correct).
We added explicit description of how the stimulus is estimated in the main text together with details about the definition of the signal perturbation. We also state clearly in the main text that the result is an estimate of the magnitude of each signal component and therefore both the identity and the intensity of the signal are estimated.
3) Decoding timevarying stimuli.
Several aspects of the analysis described in the fifth paragraph of the subsection “Frontend adaptation enhances odor decoding in complex environments” are not clear. Were odors randomly assigned to each whiff? And was the entire time course decoded, or was each whiff treated as a discrete event? In general, the description of this analysis needs to be considerably more detailed. Is there an intuitive argument as to why the longer time scales of adaptation are helpful that could be added to the last paragraph of the subsection “Frontend adaptation enhances odor decoding in complex environments”?
We have revised the text at the location indicated in the review to address the reviewer’s concerns. Note that we hold the adaptation timescale at 250 ms, but vary the memory timescale (longer ones are more helpful). We also added a phrase in the following paragraph to provide intuition as to why longer time scales of memory are useful.
4) Tests of primacy coding.
The text suggests that the background odors may interfere with primacy coding in the absence of frontend adaptation. The analysis presented in Figure 4 then shows that primacy sets are maintained in the presence of frontend adaptation. But there is not a test, unless I missed it, of the initial suggestion that primacy sets are not maintained without frontend adaptation. This test is needed to interpret this section.
We have now added on Figure 4C, the results for the nonadaptive case, showing that in absence of frontend adaptation, primacy sets are not well maintained, particularly when the primacy order is small.
5) Interplay of frontend adaptation and divisive normalization.
Figure 5 suggests that these two forms of gain control may play quite different roles in coding discrete odor categories (aversive, appetitive) and odor identity. This is quite interesting. I would consider swapping the order of presentation so you start with a discussion of odor identify (that flows more naturally from the previous sections). The differences between the present results and those of Olsen et al. should also, at a minimum, get discussed in more detail.
This is an interesting observation. We have rewritten this part of the text as suggested and have mentioned in the main text the interesting observation made by the reviewer. We have also expanded the comparison of our results with those of Olsen et al.
https://doi.org/10.7554/eLife.45293.021Article and author information
Author details
Funding
Swartz Foundation (Postdoctoral Fellowship)
 Nirag Kadakia
National Institutes of Health (R01 GM106189)
 Thierry Emonet
National Institute of Mental Health (F32 MH118700)
 Nirag Kadakia
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
Acknowledgements
NK was supported by a postdoctoral fellowship through the Swartz Foundation and by an NRSA postdoctoral fellowship through the NIH BRAIN Initiative under award number 1F32MH118700. TE was supported by NIH R01 GM106189. We thank Damon Clark, John Carlson, Mahmut Demir, Srinivas GorurShandilya, Henry Mattingly, and Ann Hermunstad for comments on the manuscript.
Senior Editor
 Catherine Dulac, Harvard University, United States
Reviewing Editor
 Fred Rieke, University of Washington, United States
Reviewer
 Katherine I. Nagel, NYU, United States
Publication history
 Received: January 23, 2019
 Accepted: June 26, 2019
 Accepted Manuscript published: June 28, 2019 (version 1)
 Version of Record published: July 4, 2019 (version 2)
Copyright
© 2019, Kadakia and Emonet
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics

 941
 Page views

 137
 Downloads

 1
 Citations
Article citation count generated by polling the highest count across the following sources: PubMed Central, Crossref, Scopus.
Download links
Downloads (link to download the article as PDF)
Download citations (links to download the citations from this article in formats compatible with various reference manager tools)
Open citations (links to open the citations from this article in various online reference manager services)
Further reading

 Computational and Systems Biology
 Developmental Biology

 Computational and Systems Biology
 Neuroscience