Abstract
Co-active or temporally ordered neural ensembles are a signature of salient sensory, motor, and cognitive events. Local convergence of such patterned activity as synaptic clusters on dendrites could help single neurons harness the potential of dendritic nonlinearities to decode neural activity patterns. We combined theory and simulations to assess the likelihood of whether projections from neural ensembles could converge onto synaptic clusters even in networks with random connectivity. Using rat hippocampal and cortical network statistics, we show that clustered convergence of axons from 3-4 different co-active ensembles is likely even in randomly connected networks, leading to representation of arbitrary input combinations in at least ten target neurons in a 100,000 population. In the presence of larger ensembles, spatiotemporally ordered convergence of 3-5 axons from temporally ordered ensembles is also likely. These active clusters result in higher neuronal activation in the presence of strong dendritic nonlinearities and low background activity. We mathematically and computationally demonstrate a tight interplay between network connectivity, spatiotemporal scales of subcellular electrical and chemical mechanisms, dendritic nonlinearities, and uncorrelated background activity. We suggest that dendritic clustered and sequence computation is pervasive, but its expression as somatic selectivity requires confluence of physiology, background activity, and connectomics.
Introduction
Neural activity patterns in the form of sequential activity or coactive ensembles reflect the structure of environmental and internal events across time, stimuli and representations. For example, sequential auditory activity may signify predators, mates, or distress calls. How can combinations of active ensembles at an input layer be discriminated by downstream neurons?
Dendrites offer a repertoire of nonlinearities that could aid neurons in achieving such computations (Ranganathan et al. 2018; K. F. H. Lee et al. 2016; Palmer et al. 2014). Several of these mechanisms can be triggered by the local convergence of inputs onto dendritic segments. For example, clustered inputs may trigger dendritic spikes (Gasparini and Magee 2006; Gasparini, Migliore, and Magee 2004). They could trigger Ca2+ release from internal stores (K. F. H. Lee et al. 2016), and other biochemical signaling cascades. Dendritic Ca2+spikes have been shown to enhance the mixing of sensory (whisker touch) and motor (whisker angle) features in the pyramidal neurons of layer 5 vibrissal cortex (Ranganathan et al. 2018). If local convergence of interesting inputs is a prerequisite for engaging such dendritic nonlinearities, a key question that remains is whether networks require specialized wiring rules to discriminate ensemble activity patterns, or whether such localized connectivity is likely to arise even with purely random connectivity.
There is considerable evidence for clustered synaptic activity in several brain regions, including the visual cortex (Gökçe, Bonhoeffer, and Scheuss 2016; Wilson et al. 2016), hippocampus (Kleindienst et al. 2011; Adoff et al. 2021) and motor cortex (Kerlin et al. 2019). A recent study done in macaques showed that synaptic inputs on dendrites are spatially clustered by stimulus features such as orientation and color (Ju et al. 2020). In the hippocampus, it has been shown that place cells receive more temporally co-active, clustered excitatory inputs within the place field of the soma than outside (Adoff et al. 2021). Notably, the length-scale of these clustered inputs is ∼10 μm, similar to the scale of operation of local Calcium-Induced Calcium Release (CICR) on the dendrite (K. F. H. Lee et al. 2016; O’Hare et al. 2022).
Many salient events are ordered in time, giving rise to convergent inputs which are not just co-active, but ordered sequentially in space and time on the dendrite. This form of dendritic computation was predicted by Rall (Rall 1964). Branco and colleagues have shown that pyramidal neurons can discriminate spatiotemporal input patterns within 100 μm dendritic zones (Branco, Clark, and Häusser 2010). (Bhalla 2017) employed reaction-diffusion models of signaling molecules to show the emergence of sequence selectivity in small dendritic zones. It is therefore valuable to analyze the statistical constraints on connectivity and input activity that can lead to synaptic clustering and dendritic sequence discrimination.
Seminal analytical work from (Poirazi and Mel 2001) showed that dendrites can increase the memory capacity of neurons by ∼23 fold by combining subsets of synaptic inputs nonlinearly. There has also been an attempt to estimate the likelihood of organization of clustered inputs using combinatorial approaches (Scheuss 2018). In the current study, we consider multiple different dimensions, including connection probability, ensemble activity, spatiotemporal scales, dendritic nonlinearities and the role of uncorrelated noise to explore the conditions under which grouped and sequential dendritic computation may emerge from random connectivity. We show that convergent connectivity of features to 3-5 synapses are statistically likely to occur on short dendritic segments in feedforward networks of ∼100,000 neurons with random connectivity. Further, these clusters may perform combinatorial input selection and sequence recognition at the dendritic level. These result in somatic selectivity under conditions of low network noise, high co-activity in upstream neurons and strong dendritic nonlinearities.
Results
We examined the mapping of network functional groupings, such as ensembles of co-active neurons onto small, <100 μm zones of synapses on neuronal dendrites. We derived approximate equations for convergence probabilities for a randomly connected feedforward network, examined the effect of noise, and then supported these estimates using connectivity based simulations. Next, we tested the effect of noise on dendritic mechanisms for sequence selectivity using electrical and chemical models. Finally, we combined these constraints to examine the likely sequence lengths that may emerge in random networks.
We framed our analysis as a two-layer feedforward network in which the input layer is organized into several ensembles within which cells are co-active in the course of representing a stimulus or other neural processing (colored dots in the presynaptic population in Figure 1A and C). In addition to neurons representing ensemble activity, the input layer also contained neurons that represented background activity/noise, whose activity was uncorrelated with the stimulus (gray dots in the presynaptic population in Figure 1C). The ‘output’ or ‘postsynaptic’ layer was where we looked for convergence of inputs. We examined the probability of projections from specific ensembles onto small local dendritic zones on a target neuron (Figure 1 A,B,C). Within this framework, we considered two cases: first, that projections are grouped within the target zone without any spatio-temporal order (grouped input). Second, that these projections are spatially ordered in the same sequence as the activity of successive ensembles (sequential input) (Figure 1 B).
The dendrite may engage different postsynaptic mechanisms at different spatiotemporal scales to mediate computation and plasticity among these inputs. We looked at three such classes of mechanisms: electrical, chemical and calcium-induced calcium release (CICR). Electrical mechanisms operate at a spatial scale of 50-100 μm and work on a faster timescale of <50ms (Golding, Staff, and Spruston 2002; Brandalise et al. 2016; Branco, Clark, and Häusser 2010; Gasparini and Magee 2006). Chemical mechanisms work over behavioral timescales of a few seconds and ∼10 μm range since they are limited by the diffusion of proteins within the dendrite (Bhalla 2017). CICR is a mechanism by which Ca2+ influx from external sources triggers the release of Ca2+ from intracellular stores thus amplifying cytosolic Ca2+ concentration. This process typically operates on a length scale of ∼10 μm and a timescale of ∼100-1000 ms (Zhou and Ross 2002; Nakamura et al. 2002). It is mediated by the inositol trisphosphate (IP3) and ryanodine receptors present on the endoplasmic reticulum. The length scales of CICR and chemical computation for grouped inputs are consistent with what are typically referred to as “synaptic clusters’’ in the field.
For our study we chose a total of six representative circuit configurations, considering hippocampal and cortical connectivity and firing statistics for each of the three above mentioned dendritic integration mechanisms (Table 1)
We used these broad categories as representatives to explore possible regimes for grouped and sequence computation in feedforward networks with random connectivity (Figure 1).
1. Grouped Convergence of Inputs
Clustered input is capable of eliciting stronger responses than the same inputs arriving in a dispersed manner (Polsky, Mel, and Schiller 2004; Schiller et al. 2000; Gasparini and Magee 2006). In principle, several electrical and signaling mechanisms with high order reactions or high cooperativity support such selectivity (Schiller et al. 2000; Kumar et al. 2018; Bhalla 2003). We illustrated this using a simple model of Calmodulin activation by Ca2+ influx (Figure 2A) from grouped and spaced synapses respectively. Grouped inputs arriving at a spacing of 2μm resulted in a higher concentration of activated calmodulin (CaM-Ca4) when compared to inputs that arrived in a dispersed manner, at a spacing of 10μm (Figure 2B). Since grouped inputs can indeed lead to larger responses at the postsynaptic dendritic branch, we investigated the likelihood of inputs converging in a grouped manner in the presence of random connectivity.
Sparse representation of fully-mixed groups of 3-4 inputs is likely in randomly connected networks
We formulated an analytical expression to estimate the bounds for the probability of convergence for grouped inputs in feedforward networks with random connectivity. Our approach is to ask how likely it is that a given set of inputs lands on a short segment of dendrite, and then scale it up to all segments on the entire dendritic length of the cell. Consider a neuron of total dendritic length L. L is in the range of 10 mm for CA1 pyramidal neurons in rats (Vitale et al. 2023; Bannister and Larkman 1995a), which we collapse into a single long notional dendrite. Our analysis does not consider electrotonic effects, hence all synapses are equivalent. Assume activity of M ensembles, each containing N neurons. We refer to groups of synapses within the target zone receiving at least one connection from each ensemble as connectivity-based fully-mixed groups (cFMG), because they hold the potential to associate information from all these ensembles (Figure 2C). We seek the probability PcFMG, that there is at least one synaptic projection from each of the M ensembles to a target zone of length Z (a group) occurring anywhere on a single target neuron of total dendritic arbor length L. The parameter p is the probability of any given neuron from the presynaptic population to connect onto a given postsynaptic cell. Values of p for the hippocampal CA3-CA1 projections have been estimated at 5% (Bolshakov and Siegelbaum 1995; Sayer, Friedlander, and Redman 1990), and for cortical circuits in the range of 5-50% (Brown and Hestrin 2009; Holmgren et al. 2003; Ko et al. 2011) (Table 1).
Expectation number of connections from any neuron in a particular ensemble, anywhere on target cell = pN
Expectation number of inputs from any neuron in a particular ensemble, on a zone of length Z on the target cell =
Probability of a zone of length Z receiving one or more inputs from a particular ensemble
Probability of a zone of length Z receiving one or more inputs from each of the M ensembles =
The probability of occurrence of at least one zone of length Z anywhere along on the dendritic arbor, such that it receives at least one input from each ensemble is given by:
Where κcFMG lies between and depending on degree of overlap across zones in different network configurations. Here σ denotes the inter-synapse interval which is 0.5 μm in all our network configurations. In Figure 2F, we compare the analytical form obtained by fitting Equation 1 with κ as the free parameter to the probabilities calculated from our connectivity based simulations (check Methods). The values of κcFMG vary for different network configurations (Supplementary Table 3)
Thus, the probability of occurrence of groups that receive connections from each of the M ensembles (PcFMG) is a function of the connection probability (p) between the two layers, the number of neurons in an ensemble (N), the relative zone-length with respect to the total dendritic arbor (Z/L) and the number of ensembles (M).
Ensemble neurons may not be 100% reliable in representing a stimulus. If we assume that each ensemble neuron is active with a probability pe during the presence of the corresponding stimulus, a connected fully-mixed group may not always receive all inputs that constitute the group in a single occurrence of the stimulus (Figure 2D). Hence the probability of a neuron receiving an active fully-mixed group where there is at least one active input arriving from each of the M ensembles is given by
Where,
and
Where κaFMG again lies between and depending on the degree of overlap between zones.
We validated these analytic expressions using simulations of network connectivity as described in the Methods (Figure 2). The simulations checked for the occurrence of such groups on each neuron in the postsynaptic population simulated through random connectivity.The simulation estimates matched the analytic expressions.
To interpret these relationships (Figure 2) we stipulate that pattern discrimination via dendritic grouped computation requires that at least one neuron in the downstream population receives convergent connections from all the M ensembles. Hence if PcFMG< 1/Tpost, where Tpost = the total number of neurons in the postsynaptic population, such projections are unlikely to happen anywhere in the network. Conversely, if PcFMG ∼1, then all the neurons in the network receive similar grouped inputs, leading to very high redundancy in representation. A sparse population code would be ideal for this scenario with a few neurons receiving such convergent grouped inputs from each pattern of ensemble activity. With ensembles of 100 neurons in the presynaptic population, we find that in a downstream population of 100,000 neurons, fully-mixed groups of 3-4 inputs are likely to occur in four out of the six network configurations we tested (hippo-elec, cortex-chem, cortex-CICR, cortex-elec) (Figure 2F). Since PaFMG is exponentially related to M, the probability of groups with greater number of inputs drops steeply as the number of ensembles increases. Large fully-mixed groups of up to 7 inputs are likely only in the cortex-electrical configuration which has higher connection probability and longer zone length. Thus grouped convergence of 3-4 inputs from a specific neural activity pattern seems likely as well as potentially non-redundant, in most network configurations.
End-effects limit convergence zones for highly branched neurons
Neurons exhibit considerable diversity with respect to their morphologies. How synapses extending across dendritic branch points interact in the context of a synaptic cluster/group, is a topic that needs detailed examination via experimental and modeling approaches. However for the sake of analysis, we present calculations under the assumption that selectivity for grouped inputs might be degraded across branch points.
Zones beginning close to a branch point might get interrupted. Consider a neuron with B branches. The length of the typical branch would be L/B. As a conservative estimate if we exclude a region of length Z for every branch, the expected number of zones that begin too close to a branch point is
For typical pyramidal neurons B∼50, so Eend ∼ 0.05 for values of Z of ∼10 µm. Thus pyramidal neurons will not be much affected by branching effects. Profusely branching neurons like Purkinje cells have B∼900 for a total L of ∼7800 µm (McConnell and Berry 1978), hence Eend ∼1 for values of Z of ∼10 µm. Thus almost all groups in Purkinje neurons would run into a branch point or terminal. For the case of electrical groups, this estimate would be scaled by a factor of 5 if we consider a zone length of 50 µm. However, it is important to note that these are very conservative estimates, as for clusters of 4-5 inputs, the number of synapses available within a zone are far greater (∼100 synapses within 50 µm).
Random connectivity supports stimulus-driven groups that mix information from different ensembles to varying extents
Since ensembles typically consist of more than one neuron which show correlated activity, this could lead to neurons receiving different combinations of ensemble inputs. For example, neurons could receive fully-mixed groups that associate inputs from 5 different ensembles {A, B, C, D, E}, or partially-mixed groups such as {A, B, A, A, D} where inputs from one or more active ensembles are missing, or homogeneous groups such as {B, B, B, B, B} which are still correlated and contain the same number of inputs, but all inputs come from a single ensemble. We refer to the superset consisting of all these different kinds of groups as stimulus-driven groups (Figure 2C).
With presynaptic ensembles of 100 neurons, stimulus-driven groups are more frequent than fully-mixed groups (compare Figure 2F with 2G and 2J with 2K; derivations in the Appendix). This gap increases further as the group size increases owing to the increasing number of permutations possible with repeating inputs (compare slopes of Figure 2F and 2G). We found that even with ensembles of 100 neurons, there is enough redundancy in input representation such that the combinatorics yield a greater proportion of partially-mixed groups, relative to fully-mixed and homogeneous groups (Figure 2L).
Background activity can also contribute to grouped activity, resulting in false positives
Networks are noisy, in the sense of considerable background activity. False positives arise when noisy inputs combine with other noisy inputs or with one or more inputs from the ensembles to trigger the same dendritic nonlinearities as true positives. We assumed the background activity to be Poisson in nature, with synapses receiving inputs at a rate R. We chose R in the range of random background activity, ∼0.1 Hz for CA3 pyramidal neurons and ∼1 Hz for the prefrontal cortex (Table 1). For chemical inputs we require strong, Ca-influx triggering bursts, which are likely to be rare. Here we assume that R is about 10% of the network background activity rate, or ∼0.01 Hz for CA3 and 0.1 Hz for cortex (Buzsáki and Mizuseki 2014; Mizuseki and Buzsáki 2013).
Assuming Poisson statistics, for a network firing at a rate of R Hz, the probability of a synapse being hit by background activity in duration D is 1-e-RD. With this we estimated the probability of occurrence of noise groups which receive grouped inputs from background activity alone (Figure 2H), and any groups which receive either background inputs or ensemble inputs or combination of both in a grouped manner (Figure 2I) (derivations in Appendix). Noise groups are more probable in the cortical configurations than hippocampal configurations owing to the higher rate of background activity. The larger time window for integration in the cortex-chem case worsens the situation as the probability for noise groups is close to one for most group sizes tested. Further, smaller noise groups (with group size of 3-5) are much more likely than larger ones.
The influence of various parameters on the probabilities of different kinds of groups is examined in Figure 2-Supplement 1 using Cortex-CICR as an example, varying one parameter at a time. Dense connectivity, longer zone length and larger ensembles all increase the probability of occurrence of fully-mixed and stimulus-driven groups.
Based on these calculations, it would be near impossible to identify neurons receiving stimulus-driven groups from other neurons in the population under high noise conditions (compare cortex-chem in Figure 2K with 2H and 2I). A comparison of the probability of active stimulus-driven groups to the probability of any groups shows that the hippo-elec configuration is the best suited for grouped computation as it has the best ‘signal-to-noise’ ratio. Hence we chose the hippo-elec as a case study below, to understand how dendritic nonlinearities influence neuronal output in our postsynaptic neurons.
Strong dendritic nonlinearities are necessary to distinguish neurons receiving stimulus-driven groups from other neurons
What significance would grouped activity on a small dendritic zone have on neuronal output, given multiple other inputs synapsing onto the dendritic arbor? The strength of dendritic nonlinearity could influence the relative contribution of grouped inputs to the output of a neuron. We evaluated how nonlinearities could distinguish neurons that receive stimulus-driven groups from other neurons in the population based on their somatic activity profiles.
We applied a nonlinear function over the number of inputs converging within a zone of length Z, using a sliding window. This provides a measure of the local activation produced due to the nonlinear combining of grouped inputs. We integrated these local activations across the entire length of the dendrite, as a proxy for somatic activation. We chose a function inspired from the cumulative of the Weibull distribution (equation shown in Figure 3A) to capture the nonlinear integration of grouped inputs (Figure 3A). This function has several interesting properties: it is smooth, nonlinear, saturates at high input numbers, and most importantly the power controls the strength of contribution that small groups have relative to the large groups. We varied the power term ρ to control the strength of the nonlinearity for the case of the hippo-elec configuration to distinguish neurons receiving groups containing 4 or more inputs.
In the presence of weak nonlinearity (ρ=2) neurons receiving grouped connectivity from stimulus-driven neurons (CSD neurons) show activation values that are similar to neurons that do not receive such groups (CSDC neurons) (Figure 3B, D). However, when the strength of the nonlinearity is increased, the CSD neurons can be easily distinguished from CSDC neurons as strong activations arise only in the presence of large groups (group size >=4) (Figure 3C, F). Under conditions of weak nonlinearity, the kinds of activation profiles seen in certain fractions of trials are also similar for CSD and CSDC neurons (Figure 3H, I). In this case, even with a more nuanced measurement than trial-averaged activation one will not be able to distinguish the two types of neurons. In contrast, in the presence of strong nonlinearity one finds a clear regime where CSD neurons show higher activation values (>0.13) in a larger fraction of trials (>0.15) (red dashed region) when compared to CSDC neurons (Figure J, K).
Even in the presence of strong nonlinearities, neurons receiving fully-mixed grouped connectivity (CFM neurons) cannot be distinguished from their complement set (CFMC neurons) (Figure 3 E, G). This is because of the large number of neurons in the CFMC group that receive either partially-mixed groups or homogeneous groups, and therefore show similar activation profiles.
In summary, we find that strong nonlinearities are necessary to distinguish the activity of neurons that receive relevant grouped connectivity from other neurons in the network. However, even in the presence of strong nonlinearities, it is not possible to distinguish neurons that receive fully-mixed groups from partially-mixed/homogeneous groups, on the basis of responses to a single pattern of ensemble activations.
2. Sequential Convergence of Inputs
Sequences of 3-5 inputs are likely, but require the presence of larger ensembles
We next performed a similar calculation with the further requirement that the ensembles are active in a certain order in time, and that the projections arrive on the target neuron in the same spatial order as their activity. This form of computation has been observed for electrical discrimination of sequences (Branco, Clark, and Häusser 2010) and has been proposed for discrimination based on chemical signaling (Bhalla 2017). Here we estimate the likelihood of the first ensemble input arriving anywhere on the dendrite, and ask how likely it is that succeeding inputs of the sequence would arrive within a set spacing. We stipulate that each successive input lies within a spacing of S to S+Δ µm with respect to the previous input. We assume that each ensemble is active over a duration D. Mean spine spacing is 0.3-1 µm in rat and mouse pyramidal neurons both in cortex and hippocampus (Bannister and Larkman 1995b; Konur et al. 2003). The typical spacing S between inputs is of the order of 2 μm for chemical/CICR sequences and ∼10 μm for electrical sequences. We assume Δ to be ∼1.5 μm for chemical/CICR and ∼5 μm for electrical sequences. Hence there can be intervening synapses between the stimulated ones.
First, we estimate the probability of connectivity-based perfectly-ordered stimulus sequences (cPOSS), which receive connections in an ordered manner from each of the ensembles. (Figure 4 A,B).
Expected number of connections from neurons in the first ensemble, anywhere on a given postsynaptic neuron = pN
Expected number of connections from neurons in the second ensemble within a spacing of S to S+Δ from a connection coming from the previous ensemble =
Expectation number of sequences containing a connection from the first ensemble followed by an connection from the second ensemble, within a spacing of S to S+Δ with respect to the first ensemble input = pNνcPOSS
Expectation number of M-length perfect sequences, that receive sequential connectivity from stimulus-driven neurons arriving from M different ensembles following the above spacing rule, on a neuron = EcPOSS = pN (νcPOSS)M − 1
Using Poisson approximation, the probability of one or more connectivity-based perfectly-ordered stimulus sequences which receive axons from M different ensembles, occurring anywhere along the dendritic length of a neuron = [Equation 4]
Thus, the probability of occurrence of sequences that receive sequential connections (PcPOSS) from each of the M ensembles is a function of the connection probability (p) between the two layers, the number of neurons in an ensemble (N), the relative window size with respect to the total dendritic arbor (Δ/L) and the number of ensembles (M).
Here we assume a specific ordering of at least one projection from each of the M ensembles. The original calculations (Bhalla 2017) for chemical sequence selectivity were symmetric, thus the ordering could be distal-to-proximal, or vice-versa. This would introduce a factor of 2 if there were no other source of symmetry breaking in the system:
For the purposes of further analysis, we assume that the neuronal system is not symmetric and hence the ordering remains unidirectional as in Equation 4.
If one considers only activity-based perfectly-ordered stimulus sequences (aPOSS) (Figure 4B) where all inputs constituting a sequence are received in a single instance of sequential stimulation of the ensembles, we get
Where , assuming that neurons in the ensemble participate with a probability of pe
When the additional constraint of order was imposed, much larger ensembles (∼1000 neurons), as opposed to 100 neurons in the case of groups, were required to obtain sequential connectivity (Figure 4D) and sequential activation (Figure 4G) of 3-5 inputs in a population of ∼100000 neurons. Hence sequence discrimination required greater redundancy in input representation in the form of larger ensembles.
Background activity can fill gaps in partial sequences to yield false positives
We next estimated the probability of occurrence of different kinds of sequences: noise sequences which receive background inputs alone (Figure 4A), any sequences which receive either background inputs or ensemble inputs or a combination of both, and gap-fill sequences which contain one more ensemble inputs that make up a partial sequence, with the missing inputs being filled in by background activity arriving in the right location at the right time (Figure 4A) (derivations in Appendix).
Similar to the case with groups, noise sequences are more probable in the cortical configurations than hippocampal configurations owing to the higher rate of background activity (Figure 4E). Cortex-chemical is again the worst affected configuration owing to the larger time window for integration. Again, smaller noise sequences of 3-5 inputs are much more likely than longer ones. We find that gap-fill sequences constitute the major fraction of sequences in most of the network configurations (compare Figure 4F, H, Figure 4-Supplement 1F). Hippo-elec again stands out as the configuration with the best signal to noise ratio (Figure 4I). Cortex-elec and hippo-CICR are the next couple of configurations that might be suitable for sequence detection.
The influence of various parameters on the probabilities of different kinds of sequences is examined in Figure 4-Supplement 1 using Cortex-CICR as an example, varying one parameter at a time. Dense connectivity, higher participation probability of ensemble neurons, longer input zone width (Δ) and larger ensembles all increase the probability of occurrence of perfectly-ordered stimulus sequences. As the ensembles get larger, gap-fills become major contributors of false-positive responses (Figure 4-Supplement 1F). Gap-fills could play a dual role. Under low noise conditions, such as in the hippo-elec configuration, gap-fills could support sequence completion in the presence of partial input. However, under high noise conditions, such as in the cortex-chem configuration, their high likelihood of occurrence can lead to very dense representations in the population, thus hampering sequence discrimination.
Our results indicate that sequential convergence of 3-5 inputs is likely with ensembles of ∼1000 neurons. Low noise conditions seem more suitable to the computation as gap-fills and noise sequences are less likely. Given the uncorrelated nature of background activity, noisy inputs arriving out of order (i.e., ectopic inputs) with respect to the sequence can influence the overall selectivity for the sequence. In the next sections we examined this question in 2 steps. First, using chemical and electrical models, we tested the effect of ectopic inputs arriving on or nearby an ongoing sequence. Next, we simulated neuronal sequence selectivity in a network including multiple patterns of ensemble stimulation, and noise, using an abstract formulation of sequence-selectivity on dendrites.
Ectopic inputs within the stimulus zone degrade selectivity for chemical sequences
A key concern with any estimate of selectivity in a network context is the arrival of extra inputs which interfere with sequence selectivity. We used computer models to estimate the effect of ectopic inputs on sequence selectivity. In this section we describe the outcomes on selectivity mediated by chemical mechanisms. We assume that the results apply both to biochemical signaling mechanisms, and to CICR, using different time-scales for the respective signaling steps.
We used a published sequence-selective reaction-diffusion system (Bhalla 2017) involving a Ca-stimulated bistable switch system with inhibitory feedback (molecule B) which slowly turns off the activated switch (molecule A) (Figure 5A). This system is highly selective for ordered/sequential stimuli in which the input arrives on successive locations spaced 3 μm apart along the simulated dendrite, at successive times (2s intervals) as compared to scrambled input in which successive locations were not stimulated in sequential order (Figure 5B). We introduced ectopic input at points within as well as flanking the zone of regular sequential input (Figure 5C). We did this for three time-points: at the start of the stimulus, in the middle, and at the end of the stimulus. In each case we assessed selectivity by comparing response to ordered input with response to a number of scrambled input patterns. Selectivity was calculated as,
Where Aseq is the response to an ordered sequence, Amean was the mean of responses to all patterns, Amax was the maximum response obtained amongst the set of patterns tested.
Our reference was the baseline selectivity to sequential input in the absence of ectopics. We observed that selectivity was strongly affected when ectopic input arrived within the regular stimulus zone, but when they were more than 2 diffusion length-constants away they had minimal effect (Figure 5D). (D= 5 μm2/s and 2 μm2/s for molecules A and B respectively).
These calculations were for a sequence length of 5. We next asked how ectopic inputs impacted selectivity for sequence lengths of M=3 to 10. We restricted our calculations to within the stimulus zone since this was where the impact of ectopic input was largest. We took the average of the selectivity for different spatiotemporal combinations of a single ectopic input within the stimulus zone, and compared it to the reference selectivity for the sequence without any other input (Figure 5E). We found that for M <5, selectivity was strongly affected (65-81% dip), but for long sequences (M >=6), selectivity recovered substantially (20-30% dip). The reduction in selectivity due to two ectopic inputs was similar to that caused by a single ectopic (Figure 5E).
Overall, ectopic inputs severely degraded selectivity for short sequences but had a smaller effect on long ones. Selectivity was not strongly affected by ectopic inputs even as close as 3 μm away from the sequence zone.
Ectopic inputs at the distal end degrade selectivity to a greater extent for electrical sequences
We next performed a similar analysis for electrical sequence selectivity, based on published experimental and modeling work from Branco et al. (Branco, Clark, and Häusser 2010). We adapted the Branco model of dendritic sequence selectivity to test the effect of ectopic input. We replicated the observation that synaptic inputs arriving in a distal-to-proximal (inward) manner at ∼2.5 μm/ms elicited a higher depolarization than those those arriving in a proximal-to-distal (outward) direction, or inputs that arrived in a scrambled order (Figure 6B). The selectivity we obtained is slightly lower than observed by Branco et al., (Branco, Clark, and Häusser 2010) since we tested a shorter sequence of 5 inputs as opposed to the longer sequence of length 9 that was used by the authors. Even though a 5-length sequence elicited relatively small EPSPs (4-6mV) at the soma, the EPSPs elicited in the dendrite were sufficiently large (20-45 mV) to engage the NMDA-mediated nonlinearity mechanism (Figure 6 - Supplement 2). As a baseline, we obtained ∼8-10% selectivity for sequences of length 5 to 9, where inputs arrived at a spacing of ∼10 μm on a 99 μm long dendrite (Figure 6D,E - reference selectivity). Selectivity was maximum at input velocities between 2-4 μm/ms, similar to what was observed experimentally (Branco, Clark, and Häusser 2010) for sequences of length 5-9 (Figure 6C). Another key distinction is that we measured selectivity as defined by Equation 6 where A represented the peak potential recorded at the soma, whereas Branco et al. (Branco, Clark, and Häusser 2010) referred to direction selectivity primarily as the difference between the responses to inward and outward input patterns.
We introduced ectopic synaptic input at different locations and times on the dendrite for ordered and scrambled sequences of 5 inputs (Figure 6D). Ectopic inputs both within and outside the zone degraded selectivity. Selectivity was most strongly disrupted by ectopic inputs arriving at the distal end. Due to the high input impedance at the distal end, ectopic inputs arriving there were not just amplifying the inward sequence response, but were boosting responses to scrambled patterns as well. The resulting narrower distribution of peak somatic EPSPs led to a further drop in selectivity (Figure 6 - Supplement 1).
We further tested the effect of ectopic inputs on sequences of different lengths. Ectopic inputs were delivered near the start, middle and end of the sequence zone, at three different times coinciding with the start, middle and end of the sequence duration. Short sequences of 3-4 inputs had low baseline selectivity, which was slightly amplified upon addition of ectopic input. With mid-length sequences of 5-7 inputs, selectivity dropped by 27-53%. Longer sequences were less affected by ectopics (16-18% drop) (Figure 7E).
In summary, ectopic inputs degraded electrical sequence selectivity even in the flanking regions, particularly when they were delivered closer to the distal end of the dendrite. Again, longer sequences were less perturbed than mid length ones.
Strong selectivity for sequences at the dendrite translates to weak selectivity at the soma
Neurons receive a wide range of inputs across their dendritic arbors. Since our previous simulations indeed showed that ectopic inputs can affect sequence selectivity, we wanted to assess how selective neurons are for sequential activity occurring on a short segment of dendrite, in the midst of all the other activity that they receive, particularly in a network scenario. In the context of a randomly connected feedforward network, can neurons receiving perfect sequential connectivity from stimulus-driven neurons (PSCSD neurons) be distinguished from other neurons based on their selectivity for sequential inputs? Since chemical sequences showed stronger reference selectivity (∼0.8) when compared to the electrical case (∼0.1) and owing to the low noise conditions, we chose the hippo-CICR configuration with 4 sequentially active ensembles to estimate sequence selectivity at the neuron level.
To test this we devised an abstract formulation to calculate a neuron’s activation based on all inputs it receives. The abstract formulation made the analysis scalable for a network context. At each location on the dendrite, the local increase in activation was a function of two things: the baseline activation contributed by an input arriving at that location and the activation buildup due to sequential inputs (Methods). The sequential contribution was scaled nonlinearly using an exponent and passed through a shifted sigmoid that bounded the maximum local activation attained within a time step. Hence sequential inputs that follow the S to S+Δ spacing rule resulted in large local activations. The local activations were integrated over time and space to obtain the overall activation of a neuron. This process was repeated for different stimulus patterns where the ensembles were activated in a different order in each pattern.
Neurons receiving sequential connectivity for the ordered stimulus [1,2,3,4] (PSCSD neurons) showed slightly higher activations on average relative to neurons that did not receive such connectivity (PSCSDC neurons) (Figure 5A,D,E). PSCSD neurons also showed higher activation to sequential stimuli than to scrambled stimuli, e.g., [3,1,2,4] (Figure 5B, G, I) or reverse stimuli [4,3,2,1] (Figure 5 C, H, I). PSCSD neurons showed slightly higher selectivity for the ordered sequence than PSCSDC neurons (Figure 5J).
Despite using a model that was tuned to show high sequence selectivity (∼0.8) at the dendrite, we observed weak somatic selectivity in the range of 0.05-0.0.08 (mean = 0.06, std = 0.01), when we accounted for all the inputs they received across their arbor. This was because the total activation of background activity and other non-sequential inputs over the entire dendritic tree overwhelmed the small signal elicited by the occurrence of a sequence. Specifically, each neuron receives ∼160 inputs from background activity over the duration of the sequence. In the presence of ensemble activity, it receives an additional ∼200 inputs scattered over the dendritic arbor. Hence a sequence of 4 inputs has a relatively small effect on this high baseline, even when it is scaled by strong nonlinearities at the local dendritic zone encompassing all the inputs in the sequence. We postulated that this problem can be alleviated when the neurons are in a balanced state, where the excitation is precisely balanced with inhibition (Hennequin, Agnes, and Vogels 2017; Bhatia, Moza, and Bhalla 2019). A previous model for sequence selectivity (Bhalla 2017), included inhibition. To nullify the effect of background activity as a proxy for the EI balanced state, we subtracted the trial averaged activation under conditions of background activity from the trial averaged activation value seen when ensembles are stimulated in addition to the background activity. This improved the selectivity to the range 0.13-0.22 (mean = 0.17, std = 0.02).
As a comparison, we ran the same simulation for the cortex-CICR configuration with 4X denser connectivity and 10X higher background activity. Under these high noise conditions, even PSCSDC show frequent occurrences of false positive (noise/gap-fill sequences). Further, after subtracting the baseline activation levels observed with background activity alone, the PSCSD neurons showed poor selectivity in the range of −0.03-0.07 (mean = 0.02, std = 0.02) (Figure 7 - Supplement 1). Hence sequence selectivity operates best under conditions of low noise.
Overall, in this section we find that even strong dendritic selectivity may yield only small somatic responses which may require additional constraints such as EI balance and low background activity in order to be resolved at the soma.
A single model exhibits multiple forms of nonlinear dendritic selectivity
We implemented all three forms of selectivity described above, in a single model which included six voltage and calcium-gated ion channels, NMDA, AMPA and GABA receptors, and chemical signaling processes in spines and dendrites. The goal of this was three fold: To show how these nonlinear operations emerge in a mechanistically detailed model, to show that they can coexist, and to show that they are separated in time-scales. We implemented a Y-branched neuron model with additional electrical compartments for the dendritic spines (Methods). This model was closely based on a published detailed chemical-electrical model (Bhalla 2017). We stimulated this model with synaptic input corresponding to the three kinds of spatiotemporal patterns described in figures Figure 8 - Supplement 1 (sequential synaptic activity triggering electrical sequence selectivity), Figure 8 - Supplement 2 (spatially grouped synaptic stimuli leading to local Ca4_CaM activation), and Figure 8 - Supplement 3 (sequential bursts of synaptic activity triggering chemical sequence selectivity). We found that each of these mechanisms show nonlinear selectivity with respect to both synaptic spacing and synaptic weights. Further, these forms of selectivity coexist in the composite model (Figure 8 - Supplements 1, 2, 3), separated by the time-scales of the stimulus patterns (∼ 100 ms, ∼ 1s and ∼10s respectively). Thus mixed signaling in active nonlinear dendrites yields selectivity of the same form as we explored in simpler individual models. A more complete analysis of the effect of morphology, branching and channel distributions deserves a separate in-depth analysis, and is outside the scope of the current study.
Discussion
We have shown using cortical and hippocampal statistics that random feedforward connectivity, along with large enough ensembles (∼100 neurons for groups and ∼1000 neurons for sequences) is sufficient to lead to the convergence of groups or sequences consisting of 3-5 inputs. This does not require further mechanisms for activity or proximity-driven axonal targeting in a population of ∼100,000 postsynaptic neurons. Such convergence provides a substrate for downstream networks to decode arbitrary input patterns.
Testing connectivity predictions using structural connectivity
Considerable advances have been made in establishing structural connectivity based on EM reconstruction (Takemura et al. 2013; W.-C. A. Lee et al. 2016; Zheng et al. 2018), viral projection-labeling (Ugolini 1995; Wickersham et al. 2007; Wall et al. 2010; Eastwood et al. 2019; Oh et al. 2014) and RNA barcoding (Chen et al. 2019; Sun et al. 2021) approaches. There have been numerous critiques of such approaches in understanding network computation (Bargmann and Marder 2013; Gomez-Marin 2021). Our analysis bridges such structural observations of connectivity and their functional implications. We focus on connectivity and activity of dendritic clusters as a potential computational motif (Pulikkottil, Somashekar, and Bhalla 2021). This focus is supported both by the known occurrence of such connectivity including functional readouts (Ju et al. 2020; Adoff et al. 2021; Kerlin et al. 2019; Fu et al. 2012), and by postsynaptic mechanisms for detecting such clustered activity (Gasparini and Magee 2006; Branco, Clark, and Häusser 2010).
Our first pass analysis shows that connectivity leading to fully-mixed conventional ‘grouped’ synaptic clusters of 3-4 inputs, is likely in feedforward networks with random connectivity (Figure 2F). Shorter groups are so dense as to occur on most neurons, which makes them hard to decode (Foldiak 2003). Further, if the uniqueness constraint is relaxed, stimulus-driven groups that receive 3-5 convergent connections from any ensemble neurons are up to two orders of magnitude more likely than fully-mixed groups (Figure 2G). Due to the additional constraint imposed by order, sequential convergence required greater redundancy in the input representations in the form of larger ensembles to give rise to similar probabilities of convergence (Figure 4).
The olfactory bulb mitral cell projections to pyramidal neurons in the piriform cortex (PC) are an example where anatomy provides predefined ensembles in the form of glomeruli, and projections that are almost random (Ghosh et al. 2011). We used our analysis to obtain insights into this circuit. In this calculation, an input ensemble consists of mitral cells receiving input from a single class of olfactory receptor neurons (ORNs). Each glomerulus is innervated by the apical tuft by ∼70 mitral/tufted cells, of which about 20 are mitral cells which project to piriform (Nagayama et al. 2014). Since each ORN class projects to two glomeruli (Ressler, Sullivan, and Buck 1994), ∼40 mitral cells are co-activated upon activation of a specific receptor class, thus forming a structurally and functionally well-defined ensemble. When odors arrive, olfactory bulb glomeruli show similar sequential activity profiles as depicted in Figure 1A (Spors and Grinvald 2002; Schaefer and Margrie 2007; Junek et al. 2010). Mitral cells predominantly project to the apical dendrites of PC neurons. These connections appear to be distributed and lack spatially structured organization within the piriform cortex (Ghosh et al. 2011; Sosulski et al. 2011; Miyamichi et al. 2011; Stettler and Axel 2009; Illig and Haberly 2003), and therefore, we treat them as randomly connected.
In mice, each PC neuron, on average, makes 0.64 synapses with neurons from a single glomerulus (Srinivasan and Stevens 2018). Thus our estimate for the expected number of inputs from a single ensemble i.e., p*N is ∼1.28 (=2*0.64). With an apical dendritic arbor of ∼2000 μm (Moreno-Velasquez et al. 2020), in a population of ∼500,000 PC neurons present in one hemisphere (Srinivasan and Stevens 2018), we predict that sparse representations of grouped mitral cell projections from 3-4 pairs of glomeruli are likely over dendritic zones ∼50 μm (PcFMG∼ 3.93*10−5, PcSDG ∼ 4.04*10−4 for group size of 4). However, convergence over the CICR/chemical length scales of 10 μm is less probable (PcFMG∼ 3.31*10−7, PcSDG ∼ 3.51*10−6 for group size of 4). κ is considered to be L/Z for these calculations. Sequential convergence over spacing windows of 10 μm is even less likely (PcPOSS∼ 4.19*10−8). Hence PC neurons are likely to receive grouped projections from 3-4 different pairs of glomeruli (representing a single odor) onto individual dendritic branches even with random connectivity. These grouped inputs could trigger NMDA spikes in the apical tuft dendrites (Kumar et al. 2018) which could aid in discriminating between odors and mediating plasticity for odor memory formation (Kumar, Barkai, and Schiller 2021).
Based on these calculations, we predict that optogenetic stimulation of combinations of three to four glomeruli should a) show clustered glutamate release on PC dendrites (can be validated using iGluSnFR (Marvin et al. 2013; 2018)), and b) show nonlinear summation as recorded from PC dendrites and possibly somas.
A similar analysis may be applicable to other neuronal circuits such as the barrel cortex and other somatosensory regions as places where the anatomy may lend itself to following up such a connectivity analysis. However, our analysis is general and applies to any circuit where computation-specific ensembles arise and connectivity is approximately random, even if they lack the convenient anatomical demarcation of ensembles provided by the olfactory bulb or somatosensory regions.
Random connectivity in conjunction with strong dendritic nonlinearities, supports mixed selectivity of 3-4 different inputs
Mixed selectivity, wherein neurons are selective to combinations of task variables as opposed to single variables, has gained a lot of interest recently (Rigotti et al. 2013; Fusi, Miller, and Rigotti 2016; Eichenbaum 2018). Neurons showing mixed selectivity have been observed in the vibrissal cortex (Ranganathan et al. 2018), hippocampus (Stefanini et al. 2020), prefrontal cortex (Kobak et al. 2016) (based on (Romo et al. 1999; Brody et al. 2003)) among other brain regions. Non-linear mixed selectivity, wherein a neuron’s activity is not a simple weighted linear combination of the different stimuli, has the potential to expand the dimensionality of neural representations (Fusi, Miller, and Rigotti 2016). Such high dimensional representations are particularly helpful in tasks involving the mixing of multiple stimuli/variables because they enable simple downstream linear readout mechanisms to decode patterns.
How do neurons achieve mixed selectivity? A possible avenue is through dendritic nonlinearities. For example, work from (Ranganathan et al. 2018) in the vibrissal cortex has shown that mixed selectivity can arise from active dendritic integration in the layer 5 pyramidal neurons, which show joint representation for sensorimotor variables. This is particularly useful in the context of adaptive sensing. Our study explores this possibility of whether dendritic computation could act as a mechanism for neurons to achieve mixed selectivity. We show that redundancy in input representation in the form of large presynaptic ensembles yields postsynaptic neurons showing different kinds of selectivities: including pure selectivity, partially-mixed and fully-mixed selectivity. In fact, it is far more likely to find neurons showing partially-mixed selectivity than neurons showing pure/fully-mixed selectivity. While such partial mixing makes it difficult to decode the occurrence of full patterns, they may still act as building blocks for neurons further downstream to aid in the decoding process.
We show that under conditions of low noise, random connectivity coupled with strong dendritic nonlinearities provides an excellent substrate for neurons to mix different sources of information. Stimulus-driven groups of 3-5 inputs are likely with ensembles of 100 neurons (Figure 2G) in 4 of the 6 network configurations tested. This mixing occurs over short segments of dendrites, enabling each neuron to potentially mix multiple different combinations of 3-5 inputs. Since partially mixed groups are more likely than fully-mixed groups, decoding the identity of patterns would require further processing based on the population response of all such neurons.
Effective pattern representation through dendritic computation needs more than structural and functional connectivity
A key prediction of our study is that structural and even functional connectivity are necessary but not sufficient conditions to achieve postsynaptic computation in clusters and sequences. Several additional factors influence the effective selectivity for clustered/sequential computation. Our analysis of the hippocampal and cortical configurations occurring over three spatiotemporal scales (electrical, CICR and chemical) indicates that background activity plays a pivotal role in determining the ability of the network to extract useful information from activity patterns. Hence we found that the hippocampal configurations are better suited for grouped as well as sequential computations, as compared to the cortex (Figure 2, 4).
Our analysis sets boundary conditions on likely postsynaptic mechanisms for dendritic cluster computations. Specifically, chemical computations occurring over timescales of several seconds are vulnerable to background inputs whose likelihood of hitting a synapse scales with the duration of the computation. (Figure 2H, I). This effect is even more pronounced in the cortex due to its higher background activity (Buzsáki and Mizuseki 2014). Hence we posit that chemical timescales are less useful for ongoing activity-driven computation leading to changes in cell firing, and rather more useful from the context of plasticity. Chemical processes occurring over seconds to minutes, can modulate synaptic weights locally within dendritic clusters, strengthening grouped inputs through cooperative plasticity (Harvey et al. 2008). Similarly, our analysis showed that electrical mechanisms worked well for grouped computations, but they were quite sensitive to ectopic inputs in the case of sequences. Further, some studies suggest that 3-5 inputs may be insufficient to trigger dendritic spikes (Gasparini, Migliore, and Magee 2004), although (Goetz, Roth, and Häusser 2021) argue that a few very strong synapses can also trigger them. CICR operating on the timescales of ∼100-1000 ms, on the other hand, could be a potential mechanism that could influence both computation and plasticity. Higher level of dendritic Ca2+ affects activity at the soma through dendritic spikes (Palmer et al. 2014) and enhances dendrite-soma correlations (O’Hare et al. 2022). In addition, Ca2+ is known to trigger downstream mechanisms that can modulate plasticity (Mateos-Aparicio and Rodríguez-Moreno 2020). Hence, CICR is a promising candidate that could boost the impact of dendritic computations in a network context.
Sequence selectivity required low background activity and strong nonlinearities to amplify sequential inputs over other kinds of inputs at the dendritic level. Chemical mechanisms showed stronger selectivity than electrical mechanisms, and they were less prone to ectopic inputs arriving outside the zone of the sequence (Figure 5, 6). However, despite using strong nonlinearities at the dendritic zone, neurons receiving perfect sequential connectivity (PSCSD) showed weak selectivity at the somatic level when we accounted for inputs arriving across the entire dendritic arbor.. The sheer number of background or non-sequential ensemble inputs impinging on the dendritic arbor led to elevated baseline activation. This problem could be alleviated through precise EI balance. If the neurons exist in the state of precise EI balance (Hennequin, Agnes, and Vogels 2017; Bhatia, Moza, and Bhalla 2019), inhibition could cancel out the effect of background activity thus amplifying the contribution of sequential inputs to the neuron’s activation.
Thus, strong nonlinearities, timescales of a few 100 ms, and low noise regimes are best suited for both grouped and sequence computations. Precise EI balance in the background inputs improves the conversion of dendritic sequence selectivity to somatic sequence selectivity.
Overall, our study provides a general framework for assessing when dendritic grouped or sequence computation can discriminate between different patterns of ensemble activations in a randomly connected feedforward network. Our approach may need further detailing to account for area-specific features of connectivity, synaptic weights, neuronal morphologies, dendritic branching patterns and network architectures. Conversely, deviations from our theory may suggest where to look for additional factors that determine connectivity and neuronal selectivity. Our analysis touches lightly on the effects of inhibition and plasticity, though these too may contribute strongly to the outcome of clustered input. While connectivity is a necessary substrate for clustered dendritic computations, we show that factors such as background activity, length and timescales of postsynaptic mechanisms, and dendritic nonlinearities can play a key role in enabling the network to tap into such computations. In addition, our analysis can guide the design of biologically-inspired networks of neurons with dendrites for solving a variety of interesting problems (Chavlis and Poirazi 2024; Iyer et al. 2022; Asabuki and Fukai 2020; Asabuki, Kokate, and Fukai 2022).
Methods
The derivations for the equations have been included in the main text and in the appendix. All simulations used python version 3.8.10. We used both analytical equations and connectivity based simulations for the results shown in Figures 3 and 5. Simulations were run using a 128 core machine running Ubuntu server 22.04.2 LTS. All code used for this paper is available at github (https://github.com/BhallaLab/synaptic-convergence-paper). The readme file in the github repository contains instructions for running the simulations.
For simulations shown in Figures 2B and 5, 8, and their corresponding supplementary figures, the MOOSE Simulator (Ray and Bhalla 2008) version 4.0.0 was used. The NEURON simulator was used for Figure 6. Version numbers of all the python packages used are available in the requirements.txt file on the github repository mentioned above.
Connectivity based simulations
For both sequences and groups, we considered a feedforward network containing a presynaptic population of Tpre neurons. Connectivity to neurons in the postsynaptic population was simulated by sampling inputs randomly from the presynaptic population onto a single long dendrite of length L, which was discretized into p*Tpre(=20000) uniformly distributed synapses. Here p is the connection probability between the two layers. In each run connectivity is sampled for 4000 neurons. Hence the connectivity to the dendrite was represented as a matrix of 4000 neurons x 20000 inputs. As simulating connectivity and examining groups/sequences for the whole postsynaptic population in a serial manner was computationally time consuming, we parallelized the process by breaking down the simulation into 100 runs consisting of 4000 postsynaptic neurons per run. This does not affect the results since the connectivity to each neuron is sampled independently, and the presynaptic activity was kept the same across different runs by initializing the same seed for generating it. However, since we wanted neurons with different connectivities, the seed for generating connectivity was changed in each run. Hence we sampled connectivity for a total of 400,000 postsynaptic neurons in all simulations except for figure 7, where we sampled 1,000,000 neurons.
Generating presynaptic activity
The first N*M neurons in the presynaptic population represented the activity of M ensembles consisting of N neurons each. The remaining neurons participated in background activity, showing Poisson firing. A total of 100 trials, consisting of 3 trial types, (connectivity based, stimulus trials, and background only trials as defined below), were simulated. In the connectivity based trial, the ensemble participation probability, pe, is set to 1, and there is no background activity. A single trial of this type was run to estimate connectivity based groups/sequences. In stimulus trials, pe was set to 0.8 and the non-ensemble neurons are considered to be firing with a rate R in a Poisson manner. Background only trials had no ensemble activity. All neurons in background-only trials were considered to be participating in background activity with a firing rate R. With the Poisson assumption, the probability of a neuron firing in duration D due to background activity was considered to be pbg= 1 - e-RD. In each trial, the activity of each presynaptic neuron was sampled from a Bernoulli distribution with p=pe or p=pbg depending on whether it was an ensemble neuron or was participating in background activity. The activity of every neuron in the presynaptic (input) population was represented as a one, minus one, or zero, where ones specified that the neuron participated in ensemble activity, minus-ones represented background activity and zeros represented no activity.
Estimating the probability of occurrence of groups
In the case of groups, neurons in all M ensembles were coactive within duration D with a probability of pe=0.8 in stimulus trials and pe=1 in connectivity based trials. For each postsynaptic neuron an input array was generated where the entry at index i represented the ensemble ID of input arriving at position i on the dendrite if the input belonged to one of the ensembles. It was −1 if the input was due to background activity, and 0 if the synapse was inactive. The input array is scanned for the occurrences of groups by sliding a window of length Z, one synapse at time. For a group of size M, if there was at least one input from each of the M ensembles within zone Z, it was considered a fully-mixed group. A stimulus-driven group was defined as one that received M or more inputs in zone Z from neurons belonging to any of the ensembles. If there were M or more inputs from background activity in a window of length Z, it was considered a noise group. We defined any group as one consisting of M or more inputs, either from the ensembles or from background activity or a combination of the two, arriving in a zone of length Z. The sum of the number of groups occurring on the dendrite was calculated for each group type (true, stimulus-driven, noise and any). The probability of a group occurring on a neuron was estimated by calculating the mean number of neurons in the postsynaptic population that received at least one group of that specific type somewhere along the dendrite.
Estimating neuronal activation for the context of grouped computation
To explore the effect of nonlinear grouped dendritic computation on neuronal response, we applied a nonlinear function on local dendritic inputs. Using the input array generated in the previous step for detecting groups, the stretched exponential function depicted in the inset in Figure 3A was applied to the total number of inputs arriving on the dendrite within a zone of length Z. In this manner, using a rolling window approach the local dendritic activation at each location on the dendrite was estimated as the output of this nonlinearity. Finally, the net activation of the neuron was measured as the summed activation across the whole dendritic arbor of length L, normalized by the total number of synapses in the zone of length Z.
Neuronal activation =
Where ω represents the nonlinear function, δ(x) = 1 if an input arrived at location x, else it is 0. L is the total dendritic length, Z the zone length and σ the inter-synapse spacing.
The nonlinearity adopted here was inspired from the cumulative distribution function of the Weibull distribution. It has the following useful properties:
It is bounded, i.e. the local activation always saturates at 1.
It is nonlinear.
By controlling the exponent, one can control the effect fewer inputs have on the activation. This is particularly relevant for tuning the contribution smaller groups have on the activation of the neuron.
One can also control the fixed value of this distribution, which is the point that remains frozen upon changing powers. We have used a value of 10 as the base of the second term, which implies that when all M inputs constituting a group are present the resulting local activation is 0.9, i.e. 90% of the maximum. When there are >M inputs, the activation saturates at a value of 1.
Estimating the probability of occurrence of sequences
A key difference between groups and sequences was that the presynaptic activity was simulated for a single time step in the case of groups. However, it was simulated over MAX_M(=9) time steps which was the maximum sequence length we considered. Hence the presynaptic activity for sequences was represented as a matrix of ones, minus ones and zeros of dimensions Tpre x 9, where ones represented ensemble activity, minus ones represented background activity and zeros represented no activity. In stimulus trials, at any given time step, only one ensemble was considered to be active. Neurons belonging to inactive ensembles in a particular time step were considered to participate in background activity in that time step. Although the presynaptic activity was simulated for a total of MAX_M time steps, while counting the number of sequences for a sequence of length M, only the activity from the first M time steps was used, as that was the length of the sequence we were interested in. Connectivity was simulated between the pre and postsynaptic neurons as described earlier.
Using this setup we looked for the occurrence of different kinds of sequences. Sequences consisted of inputs that occurred within a spacing of S to S+Δ from an input that arrived at a previous time step. Perfectly-ordered stimulus sequences (POSS) were made of ones alone. Noise sequences consisted of minus ones alone. Any sequences were formed from either ones or minus ones or a combination of both.
Since the pairwise relationships between inputs occurring in adjacent time-steps was sufficient to capture whether two inputs followed the S to S+Δ spacing, identifying synapses that followed this rule across adjacent time-steps aids in determining whether they participated in a sequence. Hence, we constructed distance matrices between synapses that followed the above spacing rule for adjacent time windows wherein the Xt->(t+1)[i,j] entry in the matrix was 1 if synapse j received input in the time window t+1, following the arrival of input at synapse i in time window t, provided synapse j occurred within a spacing of S to S+Δ of synapse i. The total number of sequences consisting of M inputs occurring between different pairs of synapses was obtained by multiplying all pairwise distance matrices and summing the entries in the product.
Total number of sequences that occurred on a neuron =
Distance matrices were constructed separately for the cases of true, noise and all sequences for each neuron and the number of different sequences in each case was calculated using the above equation. The probability of a sequence occurring on a neuron was estimated by calculating the mean number of neurons in the postsynaptic population that received at least one sequence of that specific type somewhere along the dendrite.
Estimating neuronal selectivity for the context of sequence computation
In order to estimate neuronal selectivity for ordered sequences, we stimulated the same set of ensembles in the presynaptic population in different temporal orderings, and measured the activation of neurons in the postsynaptic population. An input matrix was constructed, wherein the M[i,j]th entry was 1 if synapse j received an input at time step i. The local activation Q, at location x on the dendrite at time t was modeled as:
Here the first term captures the decay of local activation in the absence of new input, i.e. γ<1. The second term represents the baseline contribution of inputs, δ(t,x) = 1 if an input arrived at location x at time t, else it is 0. The third term represents the nonlinear contribution of sequential input. Prior activation Q within a region of Δ a distance S away from location x, is combined multiplicatively with input arriving at x at time t and is scaled nonlinearly using η. It is also passed ψ which keeps the sequential contribution within bounds.
For the analysis in Figure 7, the parameters for this formulation were chosen such that in the absence of any ectopic inputs, the local selectivity for the perfectly-ordered sequence was ∼0.8. This value is close to the value we obtained for chemical selectivity shown in Figure 5. All parameter values have been included in the supplementary.
The total activation of a neuron was obtained by integrating Q over all synapses across time to give the response Ap for each pattern p. Selectivity for the ordered sequence is measured as, Selectivity = (Aseq - Amean) / Amax
Where Aseq is the response to an ordered pattern, Amean was the mean of responses to all patterns, Amax was the maximum response obtained amongst the set of patterns tested.
Ca2+-Calmodulin simulation
With the help of MOOSE Simulator, we used the Ca2+-CaM system to model selectivity for grouped inputs. The model involves Ca2+ binding to Calmodulin in a series of steps resulting in activated Ca2+-Calmodulin (Figure 2a). Ca2+ stimulus was delivered in a spatially localized manner with synapses spaced at 2 μm, or in a dispersed manner, with inputs spaced at 10 μm on a 100 μm long dendrite which had a diameter of 5 μm. The dendrite was discretized into 1-dimensional voxels of length 0.5 μm. The mean concentration of CaM-Ca4 in the dendrite, wherein each CaM molecule is bound to 4 Ca2+ was measured over time for the clustered and dispersed input cases. The parameters of the model have been provided in the supplementary material.
Chemical simulation - Bistable-switch model
The bistable switch model was adapted from (Bhalla 2017), which was available on ModelDB at https://modeldb.science/227318. We modified the original model to include an expression for parsing ectopic inputs. The original set of parameters from (Bhalla 2017) were retained for the bistable-switch model. Simulations were performed using MOOSE. For Figure 5B, inputs were delivered either as an ordered pattern [0, 1, 2, 3, 4], or a scrambled pattern [4, 1, 0, 3, 2] with consecutive inputs spaced at a distance of 3 μm on a 100 μm long dendrite that had a diameter of 10 μm. The time interval between inputs was 2s. For Figures 6D and 6E, an additional set of 23 scrambled patterns obtained from different permutations of the perfectly-ordered sequence were also tested. Since permutations of a sequence of length M can yield a total of M! patterns, we sub-sampled from this set to select every M!/(M-4)! th pattern to get a total of 24 patterns for M>3. Within this set the first pattern was the ordered sequence itself. For the case of M=3, only the 6 available patterns (=3! permutations) were used. Response was measured as the total amount of molecule A present in the dendrite summed over time. Using these responses selectivity was calculated with the help of Equation 6.
In Figure 5D, the reference selectivity was estimated using the above formula for a set of patterns made of 5 inputs each. Further, in addition to the inputs constituting a sequence, an ectopic input was delivered. Different spatiotemporal combinations of the position and time of arrival of the ectopic input were tested. For Figure 5E, we repeated this exercise for patterns of different lengths either without any ectopic (reference), with a single ectopic or with two ectopic inputs. A smaller set of spatiotemporal pattern combinations was used in Figure 5E to keep the computational complexity within reasonable bounds. In addition, we restricted the analysis to ectopics occurring within the zone of the sequence.
Effect of ectopic inputs on electrical sequences
We chose the model published on ModelDB (https://modeldb.science/140828) for examining the effect of ectopic inputs on the selectivity of electrical sequences (Branco, Clark, and Häusser 2010). The base code from ModelDB was modified and run using the NEURON Simulator. The original channel kinetics parameters post the fix to NMDA_Mg_T initialization, and the dendrite used for testing were retained as is from the ModelDB github repository (https://github.com/ModelDBRepository/140828). We removed the timing jitter between the inputs and positioned the inputs slightly closer (∼11%) to avoid end-effects. We tested out the effects of ectopics using the passive version of the model.
For figure 6B, a sequence of 5 inputs was delivered as an inward, outward and scrambled sequence and the voltage response at the soma was measured. For figure 6C, inputs were delivered sequentially and using a number of different scrambled orderings. The voltage response at the soma was measured in each case. Selectivity was calculated using Equation 6, wherein A represented the peak excitatory postsynaptic potentials in this case. This exercise was repeated for different time intervals between successive inputs and for different sequence lengths to obtain the selectivity vs input velocity curves.
For figure 6D, E the base code was slightly modified to accommodate an ectopic input in addition to the different sequential/scrambled patterns that were being delivered. First, an ectopic input was added at different locations arriving at different times relative to the zone of the regular patterns, and selectivity is estimated at each position. The reference curve shows selectivity in the absence of the ectopic input.
This exercise was repeated for sequences of 3-9 inputs with a subset of spatio-temporal combinations of the ectopic input. We sampled a combination of 3 locations (start, middle and end) relative to the zone of the regular inputs, and 3 time points (start, middle and end) for the ectopic input. For sequences of >=6 inputs, 121 pattern combinations of the regular inputs were sampled since sampling all permutations of the sequence was computationally expensive.
Integrated multiscale model
We integrated each of the illustrated forms of spatial and temporal pattern selectivity into a single multiscale model that incorporated both electrical and chemical signaling. The model mechanisms and parameters were identical to those used in (Bhalla 2017), with a few small changes. First, this was a reduced geometry neuron with a Y-shaped dendritic structure, on which one arm of the Y was ornamented with 80 dendritic spines spaced at 2 micrometers. Second, the diffusion constant for CaM and its calcium-bound forms were set to 20 microns2/second. Third, the K_DR levels at the soma were up from 250 S/m^2 to 360 S/m^2. Fourth, the resting potential was lowered from −60 to −70 mV.
In brief, chemical signaling was present in the spines and dendrites, and these were coupled by diffusion. Chemical signaling and electrical signaling were also coupled, such that calcium influx through NMDA receptors and voltage-gated calcium channels led to signaling outcomes, and kinase activity led to channel phosphorylation which altered dendritic excitability. The model included six voltage-gated ion channels (Na, K_DR, K_A, K_Ca, K_AHP, L-type Ca) and three ligand-gated ion channels (AMPAR, NMDAR and GABAR). AMPAR and NMDAR were located on the spines only, and GABAR was located on the dendritic branches at 16 locations on each branch. Each of the 32 GABA receptors received Poisson random synaptic input at a background rate of 1.0 Hz. The detailed list of parameters is in supplementary data (Table 6, 7). In order to run the simulations for a wide range of stimulus patterns, we implemented a wrapper script around the model named allNonlin7.py. This one script was used to perform all the simulations on the integrated model using the command line options specified in the *.bat files in the github repository. Output from allNonlin7.py was saved as an hdf5 file and processed for visualization using readBackHdf7.py and analyzeNonlin6.py.
Appendix
Refer to Table 2 in the main text for parameter definitions.
Analytical derivations for different types of groups
Probability of occurrence of stimulus-driven groups
Expectation number of connections from neurons in the M ensembles, on a zone of length Z on the target neuron =
Probability of a zone of length Z receiving M or more connections from any of the neurons in the M ensembles =
Probability of a zone of length Z that receives at least M connections from any of the neurons in the M ensembles occurring anywhere along the dendritic length of a neuron,
Where κcSDG lies between and depending on the degree of overlap between zones in different network configurations.
Probability of a neuron receiving an active stimulus-driven group where there are at least M synapses receiving active inputs from neurons in the M ensembles is give by
Where
and
Where κaFMG again lies between and depending on the degree of overlap between zones.
Probability of occurrence of noise groups
Consider a network where the rate of background activity at synapses is R Hz.
Expected number of inputs arriving in duration D due to background activity at a single synapse = RD
Assuming that background activity in the presynaptic population is Poisson in nature, probability of a synapse being hit by background activity in duration D = pbg ≈ 1 − e−RD
Expected number of converging axonal connections in zone Z from the M ensembles =
Expected number of synapses being hit by noisy inputs in a dendritic zone of length Z in duration
Where σ ≈ inter-synapse interval
Probability of inputs from background activity arriving at M or more synapses in a zone of length
Probability of a zone of length Z that has at least M synapses receiving background activity in duration
Where κNG lies between and based on the degree of overlap between zones.
Probability of occurrence of any group
Expected number of synapses being hit by noisy inputs in a dendritic zone of length Z in duration
Expected number of synapses in zone Z being hit by inputs from M ensembles =
Expected number of synapses being hit either by ensemble inputs (from the M ensembles) or noisy inputs in a dendritic zone of length Z in duration
Probability of either ensemble inputs or background activity or a combination of them arriving at M or more synapses in a zone of length
Probability of a zone of length Z that has at least M synapses receiving any activity in duration
Where κAG lies between and depending on the degree of overlap between zones.
Analytical derivations for different types of sequences
Probability of occurrence of noise sequences
Assuming that background activity in the presynaptic population is Poisson in nature, probability of a synapse being hit by background activity in duration D = pbg ≈ 1 − e−RD
Expected number of synapses being hit by background activity from any neuron in the first time step D, anywhere on target neuron
Expected number of inputs from background activity arriving in the second time-step within a spacing of S to S+Δ from an input that arrived in the first time step =
Expectation number of sequences of M inputs coming from background activity in successive time steps, with the subsequent input arriving within a spacing of S to S+Δ with respect to the preceding input =
Using Poisson approximation, probability of one or more background activity driven sequences of M inputs, occurring anywhere along the dendritic length of a neuron =
Probability of occurrence of any sequence
Expected number of synapses being hit by either background activity or from activity in neurons in the first ensemble in the first time step D, anywhere on target neuron
Expected number of synapses being hit by either background activity or ensemble activity in the second time-step within a spacing of S to S+Δ from an input that arrived in the first time step =
Expected number of sequences of M inputs coming from either background activity or ensemble activity in successive time steps, with the subsequent input arriving within a spacing of S to S+Δ with respect to the preceding input =
Using Poisson approximation, probability of one or more sequences of M inputs that are driven by either background inputs or ensemble inputs or a combination of them, occurring anywhere along the dendritic length of a neuron =
Probability of occurrence of gap-fill sequences
Gap-fill sequences are those that are built from a combination of ensemble and background inputs. Hence they contain at least one input from each of the two classes.
Expectation number of M-length gap-fill sequences constructed from inputs arriving in subsequent time windows of duration D, following the S to S+Δ spacing rule, on a neuron =
Probability of occurrence of one or more gap-fill sequences containing M inputs anywhere on a neuron =
Acknowledgements
BPS and USB are at NCBS-TIFR which receives the support of the Department of Atomic Energy, Government of India, under Project Identification No. RTI 4006. We would like to acknowledge NCBS Supercomputing Facility for access to the cluster. We would like to specially thank Mukund Thattai for ideas and discussions on the project. In addition, we acknowledge Anal Kumar, Sahil Moza, Shaurya Rahul Narlanka, Sriram Narayanan, Sulu Mohan and Vinu Varghese Pulikkottil for valuable suggestions on the manuscript.
Supplementary information
Tables
Integrated multiscale model parameters and equations
Chemical model: Reduced MAPK system. All volumes are initial reference volumes, and are rescaled to actual volumes defined by the detailed morphology of the simulated neuron and subsequent spatial discretization for solving PDEs.
Electrical model: V in mV, referenced to resting potential. Time in ms.
Ion channel definitions, mostly from Traub, Wong, Miles, and Richardson. 1991. J. Neurophysiol 66:635-650.
A = normalization constant such that gGluR = gmaxGluR at peak, and
A = normalization constant such that gGABAR = gmaxGABAR at peak, and
Calcium pools:
Channel distributions
Where p = path length in microns measured along dendrite, from soma to specified point on dendritic tree, and
dia = diameter of dendrite at specified point on the dendritic tree.
Figures
References
- The Functional Organization of Excitatory Synaptic Input to Place CellsNature Communications 12https://doi.org/10.1038/s41467-021-23829-y
- Somatodendritic Consistency Check for Temporal Feature SegmentationNature Communications 11https://doi.org/10.1038/s41467-020-15367-w
- Neural Circuit Mechanisms of Hierarchical Sequence Learning Tested on Large-Scale Recording DataPLOS Computational Biology 18https://doi.org/10.1371/journal.pcbi.1010214
- Dendritic Morphology of CA1 Pyramidal Neurones from the Rat Hippocampus: I. Branching PatternsJournal of Comparative Neurology 360:150–60https://doi.org/10.1002/cne.903600111
- Dendritic Morphology of CA1 Pyramidal Neurones from the Rat Hippocampus: I. Branching PatternsJournal of Comparative Neurology 360:150–60https://doi.org/10.1002/cne.903600111
- From the Connectome to Brain FunctionNature Methods 10:483–90https://doi.org/10.1038/nmeth.2451
- Understanding Complex Signaling Networks through Models and MetaphorsProgress in Biophysics and Molecular Biology 81:45–65https://doi.org/10.1016/S0079-6107(02)00046-9
- Synaptic Input Sequence Discrimination on Behavioral Timescales Mediated by Reaction-Diffusion Chemistry in DendriteseLife 6https://doi.org/10.7554/eLife.25827
- Precise Excitation-Inhibition Balance Controls Gain and Timing in the HippocampuseLife 8https://doi.org/10.7554/eLife.43415
- Regulation of Hippocampal Transmitter Release during Development and Long-Term PotentiationScience (New York, N.Y 269:1730–34https://doi.org/10.1126/science.7569903
- Dendritic Discrimination of Temporal Input Sequences in Cortical NeuronsScience (New York, N.Y 329:1671–75https://doi.org/10.1126/science.1189664
- Dendritic NMDA Spikes Are Necessary for Timing-Dependent Associative LTP in CA3 Pyramidal CellsNature Communications 7https://doi.org/10.1038/ncomms13480
- Timing and Neural Encoding of Somatosensory Parametric Working Memory in Macaque Prefrontal CortexCerebral Cortex 13:1196–1207https://doi.org/10.1093/cercor/bhg100
- Intracortical Circuits of Pyramidal Neurons Reflect Their Long-Range Axonal TargetsNature 457:1133–36https://doi.org/10.1038/nature07658
- The Log-Dynamic Brain: How Skewed Distributions Affect Network OperationsNature Reviews Neuroscience 15:264–78https://doi.org/10.1038/nrn3687
- Dendrites Endow Artificial Neural Networks with Accurate, Robust and Parameter-Efficient LearningarXiv https://doi.org/10.48550/arXiv.2404.03708
- High-Throughput Mapping of Long-Range Neuronal Projection Using In Situ SequencingCell 179:772–786https://doi.org/10.1016/j.cell.2019.09.023
- Whole Mouse Brain Reconstruction and Registration to a Reference Atlas with Standard Histochemical Processing of Coronal SectionsJournal of Comparative Neurology 527:2170–78https://doi.org/10.1002/cne.24602
- Barlow versus Hebb: When Is It Time to Abandon the Notion of Feature Detectors and Adopt the Cell Assembly as the Unit of Cognition?Neuroscience Letters, New Perspectives on the Hippocampus and Memory 680:88–93https://doi.org/10.1016/j.neulet.2017.04.006
- Sparse Coding in the Primate CortexThe Handbook of Brain Theory and Neural Networks
- Repetitive Motor Learning Induces Coordinated Formation of Clustered Dendritic Spines in VivoNature 483:92–95https://doi.org/10.1038/nature10844
- Why Neurons Mix: High Dimensionality for Higher CognitionCurrent Opinion in Neurobiology, Neurobiology of cognitive behavior 37:66–74https://doi.org/10.1016/j.conb.2016.01.010
- State-Dependent Dendritic Computation in Hippocampal CA1 Pyramidal NeuronsJournal of Neuroscience 26:2088–2100https://doi.org/10.1523/JNEUROSCI.4428-05.2006
- On the Initiation and Propagation of Dendritic Spikes in CA1 Pyramidal NeuronsThe Journal of Neuroscience: The Official Journal of the Society for Neuroscience 24:11046–56https://doi.org/10.1523/JNEUROSCI.2520-04.2004
- Sensory Maps in the Olfactory Cortex Defined by Long-Range Viral Tracing of Single NeuronsNature 472:217–20https://doi.org/10.1038/nature09945
- Active Dendrites Enable Strong but Sparse Inputs to Determine Orientation SelectivityProceedings of the National Academy of Sciences of the United States of America 118https://doi.org/10.1073/pnas.2017339118
- Clusters of Synaptic Inputs on Dendrites of Layer 5 Pyramidal Cells in Mouse Visual CortexeLife 5https://doi.org/10.7554/eLife.09222
- Dendritic Spikes as a Mechanism for Cooperative Long-Term PotentiationNature 418:326–31https://doi.org/10.1038/nature00854
- Promisomics and the Short-Circuiting of MindeNeuro 8https://doi.org/10.1523/ENEURO.0521-20.2021
- The Spread of Ras Activity Triggered by Activation of a Single Dendritic SpineScience (New York, N.Y 321:136–40https://doi.org/10.1126/science.1159675
- Inhibitory Plasticity: Balance, Control, and CodependenceAnnual Review of Neuroscience 40:s557–79https://doi.org/10.1146/annurev-neuro-072116-031005
- Pyramidal Cell Communication within Local Networks in Layer 2/3 of Rat NeocortexThe Journal of Physiology 551https://doi.org/10.1113/jphysiol.2003.044784
- Odor-evoked activity is spatially distributed in piriform cortexJournal of Comparative Neurology 457:361–73https://doi.org/10.1002/cne.10557
- Avoiding Catastrophe: Active Dendrites Enable Multi-Task Learning in Dynamic EnvironmentsFrontiers in Neurorobotics 16https://doi.org/10.3389/fnbot.2022.846219
- Spatiotemporal Functional Organization of Excitatory Synaptic Inputs onto Macaque V1 NeuronsNature Communications 11https://doi.org/10.1038/s41467-020-14501-y
- Olfactory Coding with Patterns of Response LatenciesNeuron 67:872–84https://doi.org/10.1016/j.neuron.2010.08.005
- Functional Clustering of Dendritic Activity during Decision-MakingeLife 8https://doi.org/10.7554/eLife.46966
- Activity-Dependent Clustering of Functional Synaptic Inputs on Developing Hippocampal DendritesNeuron 72:1012–24https://doi.org/10.1016/j.neuron.2011.10.015
- Functional Specificity of Local Synaptic Connections in Neocortical NetworksNature 473:87–91https://doi.org/10.1038/nature09880
- Demixed Principal Component Analysis of Neural Population DataeLife 5https://doi.org/10.7554/eLife.10989
- Systematic Regulation of Spine Sizes and Densities in Pyramidal NeuronsJournal of Neurobiology 56:95–112https://doi.org/10.1002/neu.10229
- Plasticity of Olfactory Bulb Inputs Mediated by Dendritic NMDA-Spikes in Rodent Piriform CortexeLife 10https://doi.org/10.7554/eLife.70383
- NMDA Spikes Mediate Amplification of Inputs in the Rat Piriform CortexeLife 7https://doi.org/10.7554/eLife.38446
- Correlated Synaptic Inputs Drive Dendritic Calcium Amplification and Cooperative Plasticity during Clustered Synapse DevelopmentNeuron 89:784–99https://doi.org/10.1016/j.neuron.2016.01.012
- Anatomy and Function of an Excitatory Network in the Visual CortexNature 532:370–74https://doi.org/10.1038/nature17192
- An Optimized Fluorescent Probe for Visualizing Glutamate NeurotransmissionNature Methods 10:162–70https://doi.org/10.1038/nmeth.2333
- Stability, Affinity, and Chromatic Variants of the Glutamate Sensor iGluSnFRNature Methods 15:936–39https://doi.org/10.1038/s41592-018-0171-3
- Calcium Dynamics and Synaptic PlasticityIn Calcium Signaling Cham: Springer International Publishing :965–84https://doi.org/10.1007/978-3-030-12457-1_38
- The Effects of Undernutrition on Purkinje Cell Dendritic Growth in the RatJournal of Comparative Neurology 177:159–71https://doi.org/10.1002/cne.901770111
- Cortical Representations of Olfactory Input by Trans-Synaptic TracingNature 472:191–96https://doi.org/10.1038/nature09714
- Preconfigured, Skewed Distribution of Firing Rates in the Hippocampus and Entorhinal CortexCell Reports 4:1010–21https://doi.org/10.1016/j.celrep.2013.07.039
- Circuit-Specific Dendritic Development in the Piriform CortexeNeuro 7https://doi.org/10.1523/ENEURO.0083-20.2020
- Parallel Tufted Cell and Mitral Cell Pathways from the Olfactory Bulb to the Olfactory CortexIn The Olfactory System: From Odor Molecules to Motivational Behaviors Tokyo: Springer Japan :133–60https://doi.org/10.1007/978-4-431-54376-3_7
- Spatial Segregation and Interaction of Calcium Signalling Mechanisms in Rat Hippocampal CA1 Pyramidal NeuronsThe Journal of Physiology 543:465–80https://doi.org/10.1113/jphysiol.2002.020362
- A Mesoscale Connectome of the Mouse BrainNature 508:207–14https://doi.org/10.1038/nature13186
- Compartment-Specific Tuning of Dendritic Feature Selectivity by Intracellular Ca2+ ReleaseScience 375https://doi.org/10.1126/science.abm1670
- NMDA Spikes Enhance Action Potential Generation during Sensory InputNature Neuroscience 17:383–90https://doi.org/10.1038/nn.3646
- Impact of Active Dendrites and Structural Plasticity on the Memory Capacity of Neural TissueNeuron 29:779–96https://doi.org/10.1016/S0896-6273(01)00252-5
- Computational Subunits in Thin Dendrites of Pyramidal CellsNature Neuroscience 7:621–27https://doi.org/10.1038/nn1253
- Computation, Wiring, and Plasticity in Synaptic ClustersCurrent Opinion in Neurobiology, Computational Neuroscience 70:101–12https://doi.org/10.1016/j.conb.2021.08.001
- Theoretical Significance of Dendritic Trees for Neuronal Input-Output RelationsNeural Theory and Modeling Palo Alto: Stanford University Press https://doi.org/10.7551/mitpress/6743.003.0015
- Active Dendritic Integration and Mixed Neocortical Network Representations during an Adaptive Sensing BehaviorNature Neuroscience 21:1583–90https://doi.org/10.1038/s41593-018-0254-6
- PyMOOSE: Interoperable Scripting in Python for MOOSEFrontiers in Neuroinformatics 2
- Information Coding in the Olfactory System: Evidence for a Stereotyped and Highly Organized Epitope Map in the Olfactory BulbCell 79:1245–55https://doi.org/10.1016/0092-8674(94)90015-9
- The Importance of Mixed Selectivity in Complex Cognitive TasksNature 497:585–90https://doi.org/10.1038/nature12160
- Neuronal Correlates of Parametric Working Memory in the Prefrontal CortexNature 399:470–73https://doi.org/10.1038/20939
- The Time Course and Amplitude of EPSPs Evoked at Synapses between Pairs of CA3/CA1 Neurons in the Hippocampal SliceThe Journal of Neuroscience: The Official Journal of the Society for Neuroscience 10:826–36
- Spatiotemporal Representations in the Olfactory SystemTrends in Neurosciences 30:92–100https://doi.org/10.1016/j.tins.2007.01.001
- Quantitative Analysis of the Spatial Organization of Synaptic Inputs on the Postsynaptic DendriteFrontiers in Neural Circuits 12https://doi.org/10.3389/fncir.2018.00039
- NMDA Spikes in Basal Dendrites of Cortical Pyramidal NeuronsNature 404:285–89https://doi.org/10.1038/35005094
- Distinct Representations of Olfactory Information in Different Cortical CentresNature 472:213–16https://doi.org/10.1038/nature09868
- Spatio-Temporal Dynamics of Odor Representations in the Mammalian Olfactory BulbNeuron 34:301–15https://doi.org/10.1016/S0896-6273(02)00644-X
- The Distributed Circuit within the Piriform Cortex Makes Odor Discrimination RobustJournal of Comparative Neurology 526:2725–43https://doi.org/10.1002/cne.24492
- A Distributed Neural Code in the Dentate Gyrus and in CA1Neuron 107:703–716https://doi.org/10.1016/j.neuron.2020.05.022
- Representations of Odor in the Piriform CortexNeuron 63:854–64https://doi.org/10.1016/j.neuron.2009.09.005
- Integrating Barcoded Neuroanatomy with Spatial Transcriptional Profiling Enables Identification of Gene Correlates of ProjectionsNature Neuroscience 24:873–85https://doi.org/10.1038/s41593-021-00842-4
- A Visual Motion Detection Circuit Suggested by Drosophila ConnectomicsNature 500:175–81https://doi.org/10.1038/nature12450
- Specificity of Rabies Virus as a Transneuronal Tracer of Motor Networks: Transfer from Hypoglossal Motoneurons to Connected Second-Order and Higher Order Central Nervous System Cell GroupsJournal of Comparative Neurology 356:457–80https://doi.org/10.1002/cne.903560312
- Different Responses of Mice and Rats Hippocampus CA1 Pyramidal Neurons to in Vitro and in Vivo-like InputsFrontiers in Cellular Neuroscience 17
- Monosynaptic Circuit Tracing in Vivo through Cre-Dependent Targeting and Complementation of Modified Rabies VirusProceedings of the National Academy of Sciences 107:21848–53https://doi.org/10.1073/pnas.1011756107
- Monosynaptic Restriction of Transsynaptic Tracing from Single, Genetically Targeted NeuronsNeuron 53:639–47https://doi.org/10.1016/j.neuron.2007.01.033
- Orientation Selectivity and the Functional Clustering of Synaptic Inputs in Primary Visual CortexNature Neuroscience 19:1003–9https://doi.org/10.1038/nn.4323
- A Complete Electron Microscopy Volume of the Brain of Adult Drosophila MelanogasterCell 174:730–743https://doi.org/10.1016/j.cell.2018.06.019
- Threshold Conditions for Synaptically Evoking Ca2+Waves in Hippocampal Pyramidal NeuronsJournal of Neurophysiology 87:1799–1804https://doi.org/10.1152/jn.00601.2001
Article and author information
Author information
Version history
- Preprint posted:
- Sent for peer review:
- Reviewed Preprint version 1:
- Reviewed Preprint version 2:
Copyright
© 2024, Bhanu Priya Somashekar & Upinder Singh Bhalla
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
- views
- 287
- download
- 1
- citations
- 0
Views, downloads and citations are aggregated across all versions of this paper published by eLife.