Introduction

Neural activity patterns in the form of sequential activity or coactive ensembles reflect the structure of environmental and internal events across time, stimuli and representations. For example, sequential auditory activity may signify predators, mates, or distress calls. How can combinations of active ensembles at an input layer be discriminated by downstream neurons? This can be achieved using mixed selectivity, wherein a neuron is selective to a combination of stimulus features as opposed to showing selectivity for a single stimulus feature. Further, neurons with nonlinear mixed selectivity can increase the dimensionality of neural representations, and contribute to pattern discrimination (Rigotti et al. 2013).

Dendrites offer a repertoire of nonlinearities that could aid neurons in achieving such computations (Ranganathan et al. 2018; K. F. H. Lee et al. 2016; Palmer et al. 2014). Several of these mechanisms can be triggered by the local convergence of inputs onto dendritic segments. For example, clustered inputs may trigger dendritic spikes (Gasparini and Magee 2006; Gasparini, Migliore, and Magee 2004). They could trigger Ca2+ release from internal stores (K. F. H. Lee et al. 2016), and other biochemical signaling cascades. Dendritic Ca2+spikes have been shown to enhance the mixing of sensory (whisker touch) and motor (whisker angle) features in the pyramidal neurons of layer 5 vibrissal cortex (Ranganathan et al. 2018).

There is considerable evidence for clustered synaptic activity in several brain regions, including the visual cortex (Gökçe, Bonhoeffer, and Scheuss 2016; Wilson et al. 2016), hippocampus (Kleindienst et al. 2011; Adoff et al. 2021) and motor cortex (Kerlin et al. 2019). A recent study done in macaques showed that synaptic inputs on dendrites are spatially clustered by stimulus features such as orientation and color (Ju et al. 2020). Even though the hippocampus is known to have a non-topographic organization of place fields (Dombeck et al. 2010) and time-cells (Modi, Dhawale, and Bhalla 2014), there is evidence for clustered synaptic connections. It has been shown that place cells receive more temporally co-active, clustered excitatory inputs within the place field of the soma than outside (Adoff et al. 2021). Notably, the length-scale of these clustered inputs is ∼10 μm, similar to the scale of operation of local Calcium-Induced Calcium Release (CICR) on the dendrite (K. F. H. Lee et al. 2016). In a recent study, (O’Hare et al. 2022) augmented CICR through a knockout model and showed that it increased intra-dendritic place-field correlation which resulted in the formation of more stable place fields, again pointing to a possible role for clustered computation in dendrites.

Many salient events are ordered in time, giving rise to convergent inputs which are not just co-active, but ordered sequentially in space and time on the dendrite. This form of dendritic computation was predicted by Rall (Rall 1964). Branco and colleagues have shown that pyramidal neurons can discriminate spatiotemporal input patterns within 100 μm dendritic zones (Branco, Clark, and Häusser 2010). (Bhalla 2017) employed reaction-diffusion models of signaling molecules to show the emergence of sequence selectivity in small dendritic zones. It is therefore valuable to analyze the statistical constraints on connectivity and input activity that can lead to synaptic clustering.

Memory mechanisms are also strongly dependent on spatial details of convergence of inputs to dendrites. Using branching geometry analysis, it has been shown that the memory capacity of neurons is far greater when synapses are organized into zones which interact nonlinearly (Poirazi and Mel 2001). Memory capacity and discrimination of spatiotemporal patterns have also been shown to be a function of neuronal morphology (Zippo and Biella 2015). Many aspects of synaptic plasticity are heterosynaptic, and depend on local clustering of inputs correlated in time, reviewed in (Govindarajan, Kelleher, and Tonegawa 2006). Spatiotemporal patterns of input are known to facilitate heterosynaptic plasticity via synaptic tagging (Frey and Morris 1997) and local learning rules (Harvey and Svoboda 2007). Such cooperative dendritic plasticity rules have been proposed to result in the formation of sequential projection maps onto dendrites (Mehta 2004).

Seminal analytical work from (Poirazi and Mel 2001) showed that dendrites can increase the memory capacity of neurons by ∼23 fold by combining subsets of synaptic inputs nonlinearly. There has also been an attempt to estimate the likelihood organization of clustered inputs using combinatorial approaches (Scheuss 2018). In the current study, we consider multiple different dimensions including connection probability, ensemble activity, spatiotemporal scales, dendritic nonlinearities and the role of uncorrelated noise to explore the conditions under which grouped and sequential dendritic computation may emerge from random connectivity. We show that convergent connectivity of features to 3-5 synapses are statistically likely to occur on short dendritic segments in feedforward networks of ∼100,000 neurons with random connectivity. Further, these clusters may perform combinatorial input selection and sequence recognition at the dendritic level. These result in somatic selectivity under conditions of low network noise, high co-activity in upstream neurons and strong dendritic nonlinearities.

Results

We examined the mapping of network functional groupings, such as ensembles of co-active neurons onto small, <100 μm zones of synapses on neuronal dendrites. We derived approximate equations for convergence probabilities for a randomly connected feedforward network, examined the effect of noise, and then supported these estimates using connectivity based simulations. Next, we tested the effect of noise on dendritic mechanisms for sequence selectivity using electrical and chemical models. Finally, we combined these constraints to examine the likely sequence lengths that may emerge in random networks.

We framed our analysis as a two-layer feedforward network in which the input layer is organized into several ensembles within which cells are co-active in the course of representing a stimulus or other neural processing (colored dots in the presynaptic population in Figure 1A and C). In addition to neurons representing ensemble activity, the input layer also contained neurons that represented background activity/noise, whose activity was uncorrelated with the stimulus (gray dots in the presynaptic population in Figure 1C). The ‘output’ or ‘postsynaptic’ layer was where we looked for convergence of inputs. We examined the probability of projections from specific ensembles onto small local dendritic zones on a target neuron (Figure 1 A,B,C). Within this framework, we considered two cases: first, that projections are grouped within the target zone without any spatio-temporal order (grouped input). Second, that these projections are spatially ordered in the same sequence as the activity of successive ensembles (sequential input) (Figure 1 B).

Dendritic convergence of inputs from neural ensembles as groups and sequences.

A. Schematic of ensemble activity for grouped (left) and sequential (right) activity patterns. In this illustration there are 5 distinct ensembles represented using different colors, each with 6 neurons. In grouped activity all the ensembles are active in the same time window, whereas in sequential activity the ensembles are active successively.

B. Schematic of projection patterns of grouped (left) and sequential (right) connectivity patterns. There is no spatial organization of the grouped inputs, other than their convergence to a small target zone on the dendrite. Sequential inputs also converge to a small target zone, but connect to it in a manner such that the temporal ordering gets mapped spatially.

C. Schematic of pre and postsynaptic populations. The postsynaptic population receives connections from the presynaptic population through random connectivity. Colored dots in the presynaptic population represent ensembles, whereas gray dots represent neurons that are not part of ensembles, but may participate in background activity. Unfilled/filled dots in the presynaptic population represent neurons that are inactive/active respectively, within the temporal window of interest. The red dots in the postsynaptic population represent neurons that receive grouped/sequential connections from ensembles on their dendrites in addition to regular ensemble and background inputs, while gray dots represent those that do not receive such organized connections. Filled red dots represent neurons that receive all synaptic inputs constituting a group/sequence. Unfilled red dots are anatomically connected to a group/sequence, but some of their presynaptic neurons may be inactive in the considered time window. Filled gray dots are neurons that receive groups/sequences from background inputs either purely or in combination with some ensemble inputs, whereas unfilled gray dots represent neurons that do not receive grouped/sequential inputs. The population activity shown using filled neurons represents neurons that are active within the duration of a single occurrence of grouped/sequential activity.

The dendrite may engage different postsynaptic mechanisms at different spatiotemporal scales to mediate computation and plasticity among these inputs. We looked at three such classes of mechanisms: electrical, chemical and calcium-induced calcium release (CICR). Electrical mechanisms operate at a spatial scale of 50-100 μm and work on a faster timescale of <50ms (Golding, Staff, and Spruston 2002; Brandalise et al. 2016; Branco, Clark, and Häusser 2010; Gasparini and Magee 2006). Chemical mechanisms work over behavioral timescales of a few seconds and ∼10 μm range since they are limited by the diffusion of proteins within the dendrite (Bhalla 2017). CICR is a mechanism by which Ca2+ influx from external sources triggers the release of Ca2+ from intracellular stores thus amplifying cytosolic Ca2+ concentration. This process typically operates on a length scale of ∼10 μm and a timescale of ∼100-1000 ms (Zhou and Ross 2002; Nakamura et al. 2002). It is mediated by the inositol trisphosphate (IP3) and ryanodine receptors present on the endoplasmic reticulum. The length scales of CICR and chemical computation for grouped inputs are consistent with what are typically referred to as “synaptic clusters’’ in the field.

For our study we chose a total of six representative circuit configurations, considering hippocampal and cortical connectivity and firing statistics for each of the three above mentioned dendritic integration mechanisms (Table 1)

Example network configurations and parameters.

We used these broad categories as representatives to explore possible regimes for grouped and sequence computation in feedforward networks with random connectivity (Figure 1).

Parameters used in the analytical derivations.

1. Grouped Convergence of Inputs

Clustered input is capable of eliciting stronger responses than the same inputs arriving in a dispersed manner (Polsky, Mel, and Schiller 2004; Schiller et al. 2000; Gasparini and Magee 2006). In principle, several electrical and signaling mechanisms with high order reactions or high cooperativity support such selectivity (Schiller et al. 2000; Kumar et al. 2018; Bhalla 2003). We illustrated this using a simple model of Calmodulin activation by Ca2+ influx (Figure 2A) from grouped and spaced synapses respectively. Grouped inputs arriving at a spacing of 2μm resulted in a higher concentration of activated calmodulin (CaM-Ca4) when compared to inputs that arrived in a dispersed manner, at a spacing of 10μm (Figure 2B). Since grouped inputs can indeed lead to larger responses at the postsynaptic dendritic branch, we investigated the likelihood of inputs converging in a grouped manner in the presence of random connectivity.

Grouped convergence of noisy and stimulus-driven synaptic inputs is likely in all network configurations.

A. A simple, biologically-inspired model of Ca-Calmodulin reaction pathway.

B. The Ca-Calmodulin model shows selectivity for grouped inputs. Inputs arriving in a grouped manner (2μm spacing) lead to a higher concentration of Ca4_CaM as opposed to dispersed inputs (10μm spacing).

C. Different kinds of stimulus-driven groups. ‘Stimulus-driven groups’ receive M or more connections from neurons belonging to any of the ensembles. They could be of 3 kinds: 1) ‘fully-mixed groups’ that receive at least one connection from each ensemble; 2) ‘partially-mixed groups’ that contain multiple inputs from the same ensemble while missing inputs from others; ‘homogeneous groups’ that receive inputs from a single ensemble. D. Active and inactive groups, and noise groups. In an active group, all inputs constituting a group are received, whereas an inactive group may be missing one or more inputs in the duration of occurrence of a group, in spite of being connected with the ensembles. ‘Noise groups’ are formed of M or more background inputs; a group composed of M or more inputs of any kind, either from the ensembles or from noise or a combination of both is referred to as ‘any group’ .

E. Neurons classified as per the types of groups they receive. Note that the schematic is a qualitative representation. The sizes of the circles do not correspond to the cardinality of the sets.

F. Probability of occurrence of connectivity based fully-mixed group due to local convergence of ensemble inputs. Group size is the number of different ensembles. It also corresponds to the number of different ensembles that send axons to a group of zone length ‘Z’. The ‘CICR’ configuration overlaps with the ‘chem’ configuration of the corresponding network in F, G, J and K as they share the same zone length.

G. Probability of occurrence of connectivity based stimulus-driven group, which receives connections from any of the ensembles. Group size in G refers to the number of connections arriving from any of the ensemble neurons within zone length ‘Z’.

H. Probability of occurrence of noise group due to local noisy activation of synapses. Here the group size refers to the number of synapses activated within the zone of length ‘Z’.

I. Probability of occurrence of a group due to the convergence of either stimulus-driven inputs or noisy inputs or a combination of both. Hippo-chem overlaps with cortex-CICR as they have same value for R*D.

J, K. Probability of occurrence of active fully-mixed and active stimulus-driven groups respectively, wherein all inputs constituting a group are active. Shaded regions in F, G, H, I, J, K represent lower and upper bounds on the analytical equations for probability based on non-overlapping and overlapping cases, i.e., and in the equations .

L. Frequency distribution of groups based on the number of unique ensembles they receive connections from, for stimulus-driven groups that receive four or more ensemble inputs. Here, the total number of active ensembles in the presynaptic population is four.

M. Ratio of the probability of occurrence of an active stimulus-driven group to the probability of occurrence of any group on a neuron. This gives an indirect measure of signal to noise in the population. The mismatch seen between the analytical and simulated traces in the case of cortex-CICR is due to low sampling at higher group-sizes.

Factors affecting the likelihood of grouped convergence for different kinds of groups. Group of length 5 for the cortex-CICR configuration is chosen as the base model and each parameter is varied while keeping the others fixed.

A. Probability of groups increases with increase in connection probability between the layers in the feedforward network. Groups containing inputs from background activity are far more likely than stimulus-driven or fully-mixed groups in this case.

B. Probability of groups vs ensemble participation probability.

C. Effect of input zone length on the probability of groups. Longer zone lengths imply higher probability of group.

D. Probability of grouped convergence increases with increase in ensemble size.

E. Probability of noise group at different rates of background activity and grouped input time scales. Noise groups are more likely in networks with higher background activity and/or with mechanisms operating at slower time-scales.

F. Probability of any group at different rates of background activity and grouped input time scales. Likelihood of occurrence of any group also increases with increase in the rate of background activity and/or with mechanisms operating at slower time-scales.

Sparse representation of fully-mixed groups of 3-4 inputs is likely in randomly connected networks

We formulated an analytical expression to estimate the bounds for the probability of convergence for grouped inputs in feedforward networks with random connectivity, as follows: Consider a neuron of total dendritic length L. L is in the range of 10 mm for CA1 pyramidal neurons in rats (Vitale et al. 2023; Bannister and Larkman 1995a), which we collapse into a single long notional dendrite. Our analysis does not consider electrotonic effects, hence all synapses are equivalent. Assume activity of M ensembles, each containing N neurons. We refer to groups of synapses within the target zone receiving at least one connection from each ensemble as connectivity-based fully-mixed groups (cFMG), because they hold the potential to associate information from all these ensembles (Figure 2C). We seek the probability PcFMG, that there is at least one synaptic projection from each of the M ensembles to a target zone of length Z (a group) occurring anywhere on a single target neuron of total dendritic arbor length L. The parameter p is the probability of any given neuron from the presynaptic population to connect onto a given postsynaptic cell. Values of p for the hippocampal CA3-CA1 projections have been estimated at 5% (Bolshakov and Siegelbaum 1995; Sayer, Friedlander, and Redman 1990), and for cortical circuits in the range of 5-50% (Brown and Hestrin 2009; Holmgren et al. 2003; Ko et al. 2011) (Table 1).

Expectation number of connections from any neuron in a particular ensemble, anywhere on target cell = pN

Expectation number of inputs from any neuron in a particular ensemble, on a zone of length Z on the target cell

Probability of a zone of length Z receiving one or more inputs from a particular ensemble

Probability of a zone of length Z receiving one or more inputs from each of the M ensembles

The probability of occurrence of at least one zone of length Z anywhere along on the dendritic arbor, such that it receives at least one input from each ensemble is given by:

Where kcFMG lies between and depending on degree of overlap across zones in different network configurations. Here σ denotes the inter-synapse interval which is 0.5 μm in all our network configurations. In Figure 2F, we compare the analytical form obtained by fitting Equation 1 with k as the free parameter to the probabilities calculated from our connectivity based simulations (check Methods). The values of kcFMG vary for different network configurations (Supplementary Table 3)

Ensemble neurons may not be 100% reliable in representing a stimulus. If we assume that each ensemble neuron is active with a probability pe during the presence of the corresponding stimulus, a connected fully-mixed group may not always receive all inputs that constitute the group in a single occurrence of the stimulus (Figure 2D). Hence the probability of a neuron receiving an active fully-mixed group where there is at least one active input arriving from each of the M ensembles is given by

Where,

Where kaFMG again lies between and depending on the degree of overlap between zones.

We validated these analytic expressions using simulations of network connectivity as described in the Methods (Figure 2). The simulations checked for the occurrence of such groups on each neuron in the postsynaptic population simulated through random connectivity.The simulation estimates matched the analytic expressions.

To interpret these relationships (Figure 2) we stipulate that pattern discrimination via dendritic grouped computation requires that at least one neuron in the downstream population receives convergent connections from all the M ensembles. Hence if PcFMG< 1/Tpost, where Tpost = the total number of neurons in the postsynaptic population, such projections are unlikely to happen anywhere in the network. Conversely, if PcFMG ∼1, then all the neurons in the network receive similar grouped inputs, leading to very high redundancy in representation. A sparse population code would be ideal for this scenario with a few neurons receiving such convergent grouped inputs from each pattern of ensemble activity. With ensembles of 100 neurons in the presynaptic population, we find that in a downstream population of 100,000 neurons, fully-mixed groups of 3-4 inputs are likely to occur in four out of the six network configurations we tested (hippo-elec, cortex-chem, cortex-CICR, cortex-elec) (Figure 2F). Since PaFMG is exponentially related to M, the probability of groups with greater number of inputs drops steeply as the number of ensembles increases. Large fully-mixed groups of up to 7 inputs are likely only in the cortex-electrical configuration which has higher connection probability and longer zone length. Thus grouped convergence of 3-4 inputs from a specific neural activity pattern seems likely as well as potentially non-redundant, in most network configurations.

Random connectivity supports stimulus-driven groups that mix information from different ensembles to varying extents

Since ensembles typically consist of more than one neuron which show correlated activity, this could lead to neurons receiving different combinations of ensemble inputs. For example, neurons could receive fully-mixed groups that associate inputs from 5 different ensembles {A, B, C, D, E}, or partially-mixed groups such as {A, B, A, A, D} where inputs from one or more active ensembles are missing, or homogeneous groups such as {B, B, B, B, B} which are still correlated and contain the same number of inputs, but all inputs come from a single ensemble. We refer to the superset consisting of all these different kinds of groups as stimulus-driven groups (Figure 2C).

With presynaptic ensembles of 100 neurons, stimulus-driven groups are more frequent than fully-mixed groups (compare Figure 2F with 2G and 2J with 2K; derivations in the Appendix). This gap increases further as the group size increases owing to the increasing number of permutations possible with repeating inputs (compare slopes of Figure 2F and 2G). We found that even with ensembles of 100 neurons, there is enough redundancy in input representation such that the combinatorics yield a greater proportion of partially-mixed groups, relative to fully-mixed and homogeneous groups (Figure 2L).

Background activity can also contribute to grouped activity, resulting in false positives

Networks are noisy, in the sense of considerable background activity. False positives arise when noisy inputs combine with other noisy inputs or with one or more inputs from the ensembles to trigger the same dendritic nonlinearities as true positives. We assumed the background activity to be Poisson in nature, with synapses receiving inputs at a rate R. We chose R in the range of random background activity, ∼0.1 Hz for CA3 pyramidal neurons and ∼1 Hz for the prefrontal cortex (Table 1). For chemical inputs we require strong, Ca-influx triggering bursts, which are likely to be rare. Here we assume that R is about 10% of the network background activity rate, or ∼0.01 Hz for CA3 and 0.1 Hz for cortex (Buzsáki and Mizuseki 2014; Mizuseki and Buzsáki 2013).

Assuming Poisson statistics, for a network firing at a rate of R Hz, the probability of a synapse being hit by background activity in duration D is 1-e-RD. With this we estimated the probability of occurrence of noise groups which receive grouped inputs from background activity alone (Figure 2H), and any groups which receive either background inputs or ensemble inputs or combination of both in a grouped manner (Figure 2I) (derivations in Appendix). Noise groups are more probable in the cortical configurations than hippocampal configurations owing to the higher rate of background activity. The larger time window for integration in the cortex-chem case worsens the situation as the probability for noise groups is close to one for most group sizes tested. Further, smaller noise groups (with group size of 3-5) are much more likely than larger ones.

The influence of various parameters on the probabilities of different kinds of groups is examined in Figure 2-Supplement 1 using Cortex-CICR as an example, varying one parameter at a time. Dense connectivity, longer zone length and larger ensembles all increase the probability of occurrence of fully-mixed and stimulus-driven groups.

Based on these calculations, it would be near impossible to identify neurons receiving stimulus-driven groups from other neurons in the population under high noise conditions (compare cortex-chem in Figure 2K with 2H and 2I). A comparison of the probability of active stimulus-driven groups to the probability of any groups shows that the hippo-elec configuration is the best suited for grouped computation as it has the best ‘signal-to-noise’ ratio. Hence we chose the hippo-elec as a case study below, to understand how dendritic nonlinearities influence neuronal output in our postsynaptic neurons.

Strong dendritic nonlinearities are necessary to distinguish neurons receiving stimulus-driven groups from other neurons

What significance would grouped activity on a small dendritic zone have on neuronal output, given multiple other inputs synapsing onto the dendritic arbor? The strength of dendritic nonlinearity could influence the relative contribution of grouped inputs to the output of a neuron. We evaluated how nonlinearities could distinguish neurons that receive stimulus-driven groups from other neurons in the population based on their somatic activity profiles.

We applied a nonlinear function over the number of inputs converging within a zone of length Z, using a sliding window. This provides a measure of the local activation produced due to the nonlinear combining of grouped inputs. We integrated these local activations across the entire length of the dendrite, as a proxy for somatic activation. We chose a function inspired from the cumulative of the Weibull distribution (equation shown in Figure 3A) to capture the nonlinear integration of grouped inputs (Figure 3A). This function has several interesting properties: it is smooth, nonlinear, saturates at high input numbers, and most importantly the power controls the strength of contribution that small groups have relative to the large groups. We varied the power term ρ to control the strength of the nonlinearity for the case of the hippo-elec configuration to distinguish neurons receiving groups containing 4 or more inputs.

- Strong nonlinearities help distinguish neurons receiving stimulus-driven groups from other neurons.

A. Dendritic activation, measured as the nonlinear integration of inputs arriving within the dendritic zone Z using the formulation mentioned in the inset.

B,C. Representative total neuronal activation across 95 trials of 3 neurons that receive grouped connections from stimulus-driven neurons (CSD) (cyan) and 3 neurons that do not receive grouped connections from stimulus-driven neurons (CSDC) (violet) under conditions of weak nonlinearity (B) vs strong nonlinearity (C). Gray dots mark the occurrence of an active group of any kind.

D, E. Trial averaged neuronal activation of CSD vs CSDC neurons, and neurons receiving connectivity-based fully-mixed groups vs neurons not receiving connectivity-based fully-mixed groups, respectively, in the presence of weak nonlinearity. F, G. Similar to D, E, but in the presence of strong nonlinearity. The dashed black line in D, E, F, G represents an example threshold that would help discriminate CSD neurons from CSDC neurons based on trial averaged activation values. Strong nonlinearity helps discriminate CSD neurons from CSDC neurons. However CFM neurons cannot be distinguished from CFMC neurons even in the presence of a strong nonlinearity.

H, I, J, K. Number of neurons showing activation values within the specified range in a certain number of trials. H, J represent CSD neurons under conditions of weak and strong nonlinearities respectively. I and K represent CSDC under conditions of weak and strong nonlinearities respectively. CSDC neurons show similar activation profiles in similar fractions of trials when compared to CSD neurons. This combined with the large population size of the CSDC neurons relative to CSD neurons, makes it almost impossible to discriminate between the two groups under conditions of weak nonlinearity. In the presence of strong nonlinearities, CSD neurons show higher activation profiles (0.13 a.u. - 1.33 a.u.) in a higher percentage of trials (>15%) (the region marked with the red dashed-line). Hence they can be discriminated from CSDC neurons in the presence of a strong nonlinearity.

In the presence of weak nonlinearity (ρ=2) neurons receiving grouped connectivity from stimulus-driven neurons (CSD neurons) show activation values that are similar to neurons that do not receive such groups (CSDC neurons) (Figure 3B, D). However, when the strength of the nonlinearity is increased, the CSD neurons can be easily distinguished from CSDC neurons as strong activations arise only in the presence of large groups (group size >=4) (Figure 3C, F). Under conditions of weak nonlinearity, the kinds of activation profiles seen in certain fractions of trials are also similar for CSD and CSDC neurons (Figure 3H, I). In this case, even with a more nuanced measurement than trial-averaged activation one will not be able to distinguish the two types of neurons. In contrast, in the presence of strong nonlinearity one finds a clear regime where CSD neurons show higher activation values (>0.13) in a larger fraction of trials (>0.15) (red dashed region) when compared to CSDC neurons (Figure J, K).

Even in the presence of strong nonlinearities, neurons receiving fully-mixed grouped connectivity (CFM neurons) cannot be distinguished from their complement set (CFMC neurons) (Figure 3 E, G). This is because of the large number of neurons in the CFMC group that receive either partially-mixed groups or homogeneous groups, and therefore show similar activation profiles.

In summary, we find that strong nonlinearities are necessary to distinguish the activity of neurons that receive relevant grouped connectivity from other neurons in the network. However, even in the presence of strong nonlinearities, it is not possible to distinguish neurons that receive fully-mixed groups from partially-mixed/homogeneous groups, on the basis of responses to a single pattern of ensemble activations.

2. Sequential Convergence of Inputs

Sequences of 3-5 inputs are likely, but require the presence of larger ensembles

We next performed a similar calculation with the further requirement that the ensembles are active in a certain order in time, and that the projections arrive on the target neuron in the same spatial order as their activity. This form of computation has been observed for electrical discrimination of sequences (Branco, Clark, and Häusser 2010) and has been proposed for discrimination based on chemical signaling (Bhalla 2017). Here we stipulate that each successive input lies within a spacing of S to S+Δ µm with respect to the previous input. We assume that each ensemble is active over a duration D. Mean spine spacing is 0.3-1 µm in rat and mouse pyramidal neurons both in cortex and hippocampus (Bannister and Larkman 1995b; Konur et al. 2003). The typical spacing S between inputs is of the order of 2 μm for chemical/CICR sequences and ∼10 μm for electrical sequences. We assume Δ to be ∼1.5 μm for chemical/CICR and ∼5 μm for electrical sequences. Hence there can be intervening synapses between the stimulated ones.

First, we estimate the probability of connectivity-based perfectly-ordered stimulus sequences (cPOSS), which receive connections in an ordered manner from each of the ensembles. (Figure 4 A,B).

Sequential convergence of perfectly-ordered stimulus sequences is likely in all network configurations in the presence of larger ensembles.

A. Different kinds of sequences. True sequences are formed of ensemble inputs alone, noise sequences are formed of background inputs while gap-fills consist of a combination of ensemble and background inputs.

B. Active and inactive sequences. An active sequence receives all constituting inputs, whereas an inactive sequence may be missing one or more inputs in the duration of occurrence of a sequence, in spite of receiving axonal connections from the ensembles.

C. Neurons classified as per the types of sequences they receive. Note that the schematic is a qualitative representation. The sizes of the circles do not correspond to the cardinality of the sets.

D. Probability of occurrence of connectivity-based perfectly-ordered stimulus sequence which receives connections from all active ensembles in an ordered manner. The ‘CICR’ configuration overlaps with the ‘chem’ configuration of the corresponding network as they share the same values for S and Δ.

E. Probability of occurrence of noise sequences due to local noisy activation of synapses in an ordered fashion.

F. Probability of occurrence of gap-fill sequences that receive one or more, but not all inputs from ensembles. Background activity arriving at the right location time compensates for the inputs missing to make a perfect sequence.

G. Probability of occurrence of activity-based perfectly-ordered stimulus sequence which receives inputs from all active ensembles in an ordered manner.

H. Probability of occurrence of a sequence due to the ordered convergence of either stimulus-driven inputs or noisy inputs or a combination of both.

I. Ratio of the probability of occurrence of active perfectly-ordered stimulus sequences to the probability of occurrence of any sequence.

Factors affecting the likelihood of sequential convergence for different kinds of sequences. Sequence of length 5 for the cortex-CICR configuration is chosen as the base model and each parameter is varied while keeping the others fixed.

A. Probability of sequential convergence increases with increase in connection probability between the layers in the feedforward network for the different kinds of sequences.

B. Probability of sequences vs ensemble participation probability.

C. Effect of input zone width, Δ, on the probability of sequential convergence. Longer zone widths imply higher probability of sequence.

D. Probability of sequential convergence increases with increase in ensemble size.

E. Probability of noise sequences at different levels of background activity and sequential input time scales. Noise sequences are more likely in networks with higher background activity and/or with mechanisms operating at slower time-scales.

F. Expectation number of gap-fill sequences plotted against the expectation number of false positives (gap-fill sequences + noise sequences) for different ensemble sizes. Dot diameter scales with ensemble size. With increase in ensemble size, gap-fills constitute the major fraction of false positives.

Expected number of connections from neurons in the first ensemble, anywhere on a given postsynaptic neuron =PN

Expected number of connections from neurons in the second ensemble within a spacing of S to S+Δ from a connection coming from the previous ensemble

Expectation number of sequences containing a connection from the first ensemble followed by an connection from the second ensemble, within a spacing of S to S+Δ with respect to the first ensemble input

Expectation number of M-length perfect sequences, that receive sequential connectivity from stimulus-driven neurons arriving from M different ensembles following the above spacing rule, on a neuron

Using Poisson approximation, the probability of one or more connectivity-based perfectly-ordered stimulus sequences which receive axons from M different ensembles, occurring anywhere along the dendritic length of a neuron

Here we assume a specific ordering of at least one projection from each of the M ensembles. The original calculations (Bhalla 2017) for chemical sequence selectivity were symmetric, thus the ordering could be distal-to-proximal, or vice-versa. This would introduce a factor of 2 if there were no other source of symmetry breaking in the system:

For the purposes of further analysis, we assume that the neuronal system is not symmetric and hence the ordering remains unidirectional as in Equation 3.

If one considers only activity-based perfectly-ordered stimulus sequences (aPOSS) (Figure 4B) where all inputs constituting a sequence are received in a single instance of sequential stimulation of the ensembles, we get

Where , assuming that neurons in the ensemble participate with a probability of pe

When the additional constraint of order was imposed, much larger ensembles (∼1000 neurons), as opposed to 100 neurons in the case of groups, were required to obtain sequential connectivity (Figure 4D) and sequential activation (Figure 4G) of 3-5 inputs in a population of ∼100000 neurons. Hence sequence discrimination required greater redundancy in input representation in the form of larger ensembles.

Background activity can fill gaps in partial sequences to yield false positives

We next estimated the probability of occurrence of different kinds of sequences: noise sequences which receive background inputs alone (Figure 4A), any sequences which receive either background inputs or ensemble inputs or a combination of both, and gap-fill sequences which contain one more ensemble inputs that make up a partial sequence, with the missing inputs being filled in by background activity arriving in the right location at the right time (Figure 4A) (derivations in Appendix).

Similar to the case with groups, noise sequences are more probable in the cortical configurations than hippocampal configurations owing to the higher rate of background activity (Figure 4E). Cortex-chemical is again the worst affected configuration owing to the larger time window for integration. Again, smaller noise sequences of 3-5 inputs are much more likely than longer ones. We find that gap-fill sequences constitute the major fraction of sequences in most of the network configurations (compare Figure 4F, H, Figure 4-Supplement 1F). Hippo-elec again stands out as the configuration with the best signal to noise ratio (Figure 4I). Cortex-elec and hippo-CICR are the next couple of configurations that might be suitable for sequence detection.

The influence of various parameters on the probabilities of different kinds of sequences is examined in Figure 4-Supplement 1 using Cortex-CICR as an example, varying one parameter at a time. Dense connectivity, higher participation probability of ensemble neurons, longer input zone width (Δ) and larger ensembles all increase the probability of occurrence of perfectly-ordered stimulus sequences. As the ensembles get larger, gap-fills become major contributors of false-positive responses (Figure 4-Supplement 1F). Gap-fills could play a dual role. Under low noise conditions, such as in the hippo-elec configuration, gap-fills could support sequence completion in the presence of partial input. However, under high noise conditions, such as in the cortex-chem configuration, their high likelihood of occurrence can lead to very dense representations in the population, thus hampering sequence discrimination.

Our results indicate that sequential convergence of 3-5 inputs is likely with ensembles of ∼1000 neurons. Low noise conditions seem more suitable to the computation as gap-fills and noise sequences are less likely. Given the uncorrelated nature of background activity, noisy inputs arriving out of order (i.e., ectopic inputs) with respect to the sequence can influence the overall selectivity for the sequence. In the next sections we examined this question in 2 steps. First, using chemical and electrical models, we tested the effect of ectopic inputs arriving on or nearby an ongoing sequence. Next, we simulated neuronal sequence selectivity in a network including multiple patterns of ensemble stimulation, and noise, using an abstract formulation of sequence-selectivity on dendrites.

Ectopic inputs within the stimulus zone degrade selectivity for chemical sequences

A key concern with any estimate of selectivity in a network context is the arrival of extra inputs which interfere with sequence selectivity. We used computer models to estimate the effect of ectopic inputs on sequence selectivity. In this section we describe the outcomes on selectivity mediated by chemical mechanisms. We assume that the results apply both to biochemical signaling mechanisms, and to CICR, using different time-scales for the respective signaling steps.

We used a published sequence-selective reaction-diffusion system (Bhalla 2017) involving a Ca-stimulated bistable switch system with inhibitory feedback (molecule B) which slowly turns off the activated switch (molecule A) (Figure 5A). This system is highly selective for ordered/sequential stimuli in which the input arrives on successive locations spaced 3 μm apart along the simulated dendrite, at successive times (2s intervals) as compared to scrambled input in which successive locations were not stimulated in sequential order (Figure 5B). We introduced ectopic input at points within as well as flanking the zone of regular sequential input (Figure 5C). We did this for three time-points: at the start of the stimulus, in the middle, and at the end of the stimulus. In each case we assessed selectivity by comparing response to ordered input with response to a number of scrambled input patterns. Selectivity was calculated as,

Ectopic input to chemical reaction-diffusion sequence selective system.

A. Schematic of chemical system. Calcium activates molecule A, which activates itself, and has negative inhibitory feedback involving molecule B.

B. Snapshot of concentration of molecule A in a one-dimensional diffusion system, following sequential (blue) and scrambled (orange) stimuli (methods).

C. Geometry for analysis of influence of ectopic stimulation. Colored arrows represent the regular input for sequential/scrambled patterns. Gray arrows represent positions where the effect of ectopic inputs was tested.

D. Selectivity as a function of location of ectopic input. Ectopic input was given at three times: the start of the regular stimulus, the middle of the stimulus, and at the stimulus end. Cyan shaded region marks the zone of regular inputs. Ectopic stimuli were given at one μm spacing intervals to capture any steep changes in selectivity. The reference line indicates the baseline selectivity in the absence of any ectopic input. E. Selectivity as a function of sequence length and number of ectopic inputs.

Where Aseq is the response to an ordered sequence, Amean was the mean of responses to all patterns, Amax was the maximum response obtained amongst the set of patterns tested.

Our reference was the baseline selectivity to sequential input in the absence of ectopics. We observed that selectivity was strongly affected when ectopic input arrived within the regular stimulus zone, but when they were more than 2 diffusion length-constants away they had minimal effect (Figure 5D). (D= 5 μm2/s and 2 μm2/s for molecules A and B respectively).

These calculations were for a sequence length of 5. We next asked how ectopic inputs impacted selectivity for sequence lengths of M=3 to 10. We restricted our calculations to within the stimulus zone since this was where the impact of ectopic input was largest. We took the average of the selectivity for different spatiotemporal combinations of a single ectopic input within the stimulus zone, and compared it to the reference selectivity for the sequence without any other input (Figure 5E). We found that for M <5, selectivity was strongly affected (65-81% dip), but for long sequences (M >=6), selectivity recovered substantially (20-30% dip). The reduction in selectivity due to two ectopic inputs was similar to that caused by a single ectopic (Figure 5E).

Overall, ectopic inputs severely degraded selectivity for short sequences but had a smaller effect on long ones. Selectivity was not strongly affected by ectopic inputs even as close as 3 μm away from the sequence zone.

Ectopic inputs at the distal end degrade selectivity to a greater extent for electrical sequences

We next performed a similar analysis for electrical sequence selectivity, based on published experimental and modeling work from Branco et al. (Branco, Clark, and Häusser 2010). We adapted the Branco model of dendritic sequence selectivity to test the effect of ectopic input. We replicated the observation that synaptic inputs arriving in a distal-to-proximal (inward) manner at ∼2.5 μm/ms elicited a higher depolarization than those those arriving in a proximal-to-distal (outward) direction, or inputs that arrived in a scrambled order (Figure 6B). The selectivity we obtained is slightly lower than observed by Branco et al., (Branco, Clark, and Häusser 2010) since we tested a shorter sequence of 5 inputs as opposed to the longer sequence of length 9 that was used by the authors. As a baseline, we obtained ∼8-10% selectivity for sequences of length 5 to 9, where inputs arrived at a spacing of ∼10 μm on a 99 μm long dendrite (Figure 6D,E - reference selectivity). Selectivity was maximum at input velocities between 2-4 μm/ms, similar to what was observed experimentally (Branco, Clark, and Häusser 2010) for sequences of length 5-9 (Figure 6C). Another key distinction is that we measured selectivity as defined by Equation 5 where A represented the peak potential recorded at the soma, whereas Branco et al. (Branco, Clark, and Häusser 2010) referred to direction selectivity primarily as the difference between the responses to inward and outward input patterns.

Effect of ectopic input on electrical sequence selectivity.

A: Neuronal morphology used for examining the effect of ectopic input in an electrical model. The selected dendrite for examining selectivity has been indicated in magenta.

B. Soma response to three different input patterns containing 5 inputs: perfectly-ordered inward sequence, outward sequence and scrambled pattern.

C. Dependence of selectivity on stimulus speed, that is, how fast successive inputs were delivered for a sequence of 5 inputs.

D. Effect of ectopic input at different stimulus locations along the dendrite, for sequence of length 5. Shaded cyan region indicates the location of the regular synaptic input. Stimuli were given at time t_start: the start of the regular stimulus sequence; t_mid: the middle of the sequence, and t_end: the end of the sequence.

E: Effect of single ectopic input as a function of sequence length. Selectivity drops when there is a single ectopic input for sequences with >=5 inputs.

Strong disruption of selectivity at the distal end is caused by decrease in variance among the responses to different patterns.

A. Same as Figure 6D. Effect of a single ectopic input on sequence selectivity at different locations along the dendrite.

B,C - The distribution of responses of different input patterns in the absence (reference) and presence of ectopic input arriving at different time points (t_start, t_mid, t_end). The inverted triangle symbol marks the response to the inward sequence. B is a violin plot representation, whereas C depicts the same information as a swarm plot.

We introduced ectopic synaptic input at different locations and times on the dendrite for ordered and scrambled sequences of 5 inputs (Figure 6D). Ectopic inputs both within and outside the zone degraded selectivity. Selectivity was most strongly disrupted by ectopic inputs arriving at the distal end. Due to the high input impedance at the distal end, ectopic inputs arriving there were not just amplifying the inward sequence response, but were boosting responses to scrambled patterns as well. The resulting narrower distribution of peak somatic EPSPs led to a further drop in selectivity (Figure 6 - Supplement 1).

We further tested the effect of ectopic inputs on sequences of different lengths. Ectopic inputs were delivered near the start, middle and end of the sequence zone, at three different times coinciding with the start, middle and end of the sequence duration. Short sequences of 3-4 inputs had low baseline selectivity, which was slightly amplified upon addition of ectopic input. With mid-length sequences of 5-7 inputs, selectivity dropped by 27-53%. Longer sequences were less affected by ectopics (16-18% drop) (Figure 7E).

Neurons receiving perfectly-ordered stimulus sequences show weak selectivity.

A, B, C. Representative total neuronal activation across 48 trials of 3 neurons that receive perfect sequential connections from stimulus-driven neurons (PSCSD) (red) and 3 neurons that do not receive such connections from stimulus-driven neurons (PSCSDC) (gray) for three stimulus patterns: Perfect ordered activation of ensembles (A), scrambled activation (B) and activation of ensembles in the reverse order (C). Gray dots mark the occurrence of any sequence.

D, E. Representative activation values across trials for pattern #1 for the first neuron in each group: the PSCSD group shown in red (D) and the PSCSDC group shown in gray (E). Neurons are riding on high baseline activation, with small fluctuations when a complete active sequence occurs.

F, G, H. Trial averaged neuronal activation of PSCSD vs PSCSDC neurons, for the three stimulus patterns indicated in A, B, C respectively.

I. Trial averaged neuronal activation of neurons classified as PSCSD vs PSCSDC in the presence of background activity alone, i.e., when the ensembles are not stimulated.

J. Trial averaged neuronal activation of 19 PSCSD neurons (depicted in different colors) in response to 24 different stimulus patterns. Pattern #1 corresponds to the ordered sequence, while pattern #24 is the reverse sequence.

K. Sequence selectivity of PSCSD and PSCSDC neurons, without (one-sided Mann-Whitney U Test, p=2.18*10-14) and with background activity subtraction (one-sided Mann-Whitney U Test, p=2.18*10-14).

Neurons receiving perfect stimulus driven sequences show poor selectivity in the presence of high background activity, such as the cortex-CICR configuration.

A, B, C. Representative total neuronal activation across 48 trials of 3 neurons that receive perfect sequential connections from stimulus-driven neurons (PSCSD) (red) and 3 neurons that do not receive such connections from stimulus-driven neurons (PSCSDC) (gray) for three stimulus patterns: Perfect ordered activation of ensembles (A), scrambled activation (B) and activation of ensembles in the reverse order (C). Gray dots mark the occurrence of any sequence.

D, E. Representative activation values across trials for pattern #1 for the first neuron in each group: the PSCSD group shown in red (D) and the PSCSDC group shown in gray (E). Neurons are riding on high baseline activation, with small fluctuations when a complete active sequence occurs.

F, G, H. Trial averaged neuronal activation of PSCSD vs PSCSDC neurons, for the three stimulus patterns indicated in A, B, C respectively.

I. Trial averaged neuronal activation of neurons classified as PSCSD vs PSCSDC in the presence of background activity alone, i.e., when the ensembles are not stimulated.

J. Trial averaged neuronal activation of 20 PSCSD neurons (depicted in different colors) in response to 24 different stimulus patterns. Pattern #1 corresponds to the ordered sequence, while pattern #24 is the reverse sequence.

K. Sequence selectivity of PSCSD and PSCSDC neurons, without and with background activity subtraction.

In summary, ectopic inputs degraded electrical sequence selectivity even in the flanking regions, particularly when they were delivered closer to the distal end of the dendrite. Again, longer sequences were less perturbed than mid length ones.

Strong selectivity for sequences at the dendrite translates to weak selectivity at the soma

Neurons receive a wide range of inputs across their dendritic arbors. We wanted to assess how selective neurons are for sequential activity occurring on a short segment of dendrite, in the midst of all the other activity that they receive. In the context of a randomly connected feedforward network, can neurons receiving perfect sequential connectivity from stimulus-driven neurons (PSCSD neurons) be distinguished from other neurons based on their selectivity for sequential inputs? Since chemical sequences showed stronger reference selectivity (∼0.8) when compared to the electrical case (∼0.1) and owing to the low noise conditions, we chose the hippo-CICR configuration with 4 sequentially active ensembles to estimate sequence selectivity at the neuron level.

To test this we devised an abstract formulation to calculate a neuron’s activation based on all inputs it receives. At each location on the dendrite, the local increase in activation was a function of two things: the baseline activation contributed by an input arriving at that location and the activation buildup due to sequential inputs (Methods). The sequential contribution was scaled nonlinearly using an exponent and passed through a shifted sigmoid that bounded the maximum local activation attained within a time step. Hence sequential inputs that follow the S to S+Δ spacing rule resulted in large local activations. The local activations were integrated over time and space to obtain the overall activation of a neuron. This process was repeated for different stimulus patterns where the ensembles were activated in a different order in each pattern.

Neurons receiving sequential connectivity for the ordered stimulus [1,2,3,4] (PSCSD neurons) showed slightly higher activations on average relative to neurons that did not receive such connectivity (PSCSDC neurons) (Figure 5A,D,E). PSCSD neurons also showed higher activation to sequential stimuli than to scrambled stimuli, e.g., [3,1,2,4] (Figure 5B, G, I) or reverse stimuli [4,3,2,1] (Figure 5 C, H, I). PSCSD neurons showed slightly higher selectivity for the ordered sequence than PSCSDC neurons (Figure 5J).

Despite using a model that was tuned to show high sequence selectivity (∼0.8) at the dendrite, we observed weak somatic selectivity in the range of 0.05-0.0.08 (mean = 0.06, std = 0.01), when we accounted for all the inputs they received across their arbor. This was because the total activation of background activity and other non-sequential inputs over the entire dendritic tree overwhelmed the small signal elicited by the occurrence of a sequence. Specifically, each neuron receives ∼160 inputs from background activity over the duration of the sequence. In the presence of ensemble activity, it receives an additional ∼200 inputs scattered over the dendritic arbor. Hence a sequence of 4 inputs has a relatively small effect on this high baseline, even when it is scaled by strong nonlinearities at the local dendritic zone encompassing all the inputs in the sequence. We postulated that this problem can be alleviated when the neurons are in a balanced state, where the excitation is precisely balanced with inhibition (Hennequin, Agnes, and Vogels 2017; Bhatia, Moza, and Bhalla 2019). A previous model for sequence selectivity (Bhalla 2017), included inhibition. To nullify the effect of background activity as a proxy for the EI balanced state, we subtracted the trial averaged activation under conditions of background activity from the trial averaged activation value seen when ensembles are stimulated in addition to the background activity. This improved the selectivity to the range 0.13-0.22 (mean = 0.17, std = 0.02) .

As a comparison, we ran the same simulation for the cortex-CICR configuration with 4X denser connectivity and 10X higher background activity. Under these high noise conditions, even PSCSDC show frequent occurrences of false positive (noise/gap-fill sequences). Further, after subtracting the baseline activation levels observed with background activity alone, the PSCSD neurons showed poor selectivity in the range of −0.03-0.07 (mean = 0.02, std = 0.02) (Figure 7 - Supplement 1). Hence sequence selectivity operates best under conditions of low noise.

Overall, in this section we find that even strong dendritic selectivity may yield only small somatic responses which may require additional constraints such as EI balance and low background activity in order to be resolved at the soma.

Discussion

We have shown using cortical and hippocampal statistics that random feedforward connectivity, along with large enough ensembles (∼100 neurons for groups and ∼1000 neurons for sequences) is sufficient to lead to the convergence of groups or sequences consisting of 3-5 inputs. This does not require further mechanisms for activity or proximity-driven axonal targeting in a population of ∼100,000 postsynaptic neurons. Such convergence provides a substrate for downstream networks to decode arbitrary input patterns.

Testing connectivity predictions using structural connectivity

Considerable advances have been made in establishing structural connectivity based on EM reconstruction (Takemura et al. 2013; W.-C. A. Lee et al. 2016; Zheng et al. 2018), viral projection-labeling (Ugolini 1995; Wickersham et al. 2007; Wall et al. 2010; Eastwood et al. 2019; Oh et al. 2014) and RNA barcoding (Chen et al. 2019; Sun et al. 2021) approaches. There have been numerous critiques of such approaches in understanding network computation (Bargmann and Marder 2013; Gomez-Marin 2021). Our analysis bridges such structural observations of connectivity and their functional implications. We focus on connectivity and activity of dendritic clusters as a potential computational motif (Pulikkottil, Somashekar, and Bhalla 2021). This focus is supported both by the known occurrence of such connectivity including functional readouts (Ju et al. 2020; Adoff et al. 2021; Kerlin et al. 2019; Fu et al. 2012), and by postsynaptic mechanisms for detecting such clustered activity (Gasparini and Magee 2006; Branco, Clark, and Häusser 2010).

Our first pass analysis shows that connectivity leading to fully-mixed conventional ‘grouped’ synaptic clusters of 3-4 inputs, is likely in feedforward networks with random connectivity (Figure 2F). Shorter groups are so dense as to occur on most neurons, which makes them hard to decode (Foldiak 2003). Further, if the uniqueness constraint is relaxed, stimulus-driven groups that receive 3-5 convergent connections from any ensemble neurons are up to two orders of magnitude more likely than fully-mixed groups (Figure 2G). Due to the additional constraint imposed by order, sequential convergence required greater redundancy in the input representations in the form of larger ensembles to give rise to similar probabilities of convergence (Figure 4).

The olfactory bulb mitral cell projections to pyramidal neurons in the piriform cortex (PC) are an example where anatomy provides predefined ensembles in the form of glomeruli, and projections that are almost random (Ghosh et al. 2011). We used our analysis to obtain insights into this circuit. In this calculation, an input ensemble consists of mitral cells receiving input from a single class of olfactory receptor neurons (ORNs). Each glomerulus is innervated by the apical tuft by ∼70 mitral/tufted cells, of which about 20 are mitral cells which project to piriform (Nagayama et al. 2014). Since each ORN class projects to two glomeruli (Ressler, Sullivan, and Buck 1994), ∼40 mitral cells are co-activated upon activation of a specific receptor class, thus forming a structurally and functionally well-defined ensemble. When odors arrive, olfactory bulb glomeruli show similar sequential activity profiles as depicted in Figure 1A (Spors and Grinvald 2002; Schaefer and Margrie 2007; Junek et al. 2010). Mitral cells predominantly project to the apical dendrites of PC neurons. These connections appear to be distributed and lack spatially structured organization within the piriform cortex (Ghosh et al. 2011; Sosulski et al. 2011; Miyamichi et al. 2011; Stettler and Axel 2009; Illig and Haberly 2003), and therefore, we treat them as randomly connected.

In mice, each PC neuron, on average, makes 0.64 synapses with neurons from a single glomerulus (Srinivasan and Stevens 2018). Thus our estimate for the expected number of inputs from a single ensemble i.e., p*N is ∼1.28 (=2*0.64). With an apical dendritic arbor of ∼2000 μm (Moreno-Velasquez et al. 2020), in a population of ∼500,000 PC neurons present in one hemisphere (Srinivasan and Stevens 2018), we predict that sparse representations of grouped mitral cell projections from 3-4 pairs of glomeruli are likely over dendritic zones ∼50 μm (PcFMG∼ 3.93*10-5, PcSDG ∼ 4.04*10-4 for group size of 4). However, convergence over the CICR/chemical length scales of 10 μm is less probable (PcFMG∼ 3.31*10-7, PcSDG ∼ 3.51*10-6 for group size of 4). k is considered to be L/Z for these calculations. Sequential convergence over spacing windows of 10 μm is even less likely (PcPOSS∼ 4.19*10-8). Hence PC neurons are likely to receive grouped projections from 3-4 different pairs of glomeruli (representing a single odor) onto individual dendritic branches even with random connectivity. These grouped inputs could trigger NMDA spikes in the apical tuft dendrites (Kumar et al. 2018) which could aid in discriminating between odors and mediating plasticity for odor memory formation (Kumar, Barkai, and Schiller 2021).

Based on these calculations, we predict that optogenetic stimulation of combinations of three to four glomeruli should a) show clustered glutamate release on PC dendrites (can be validated using iGluSnFR (Marvin et al. 2013; 2018)), and b) show nonlinear summation as recorded from PC dendrites and possibly somas.

A similar analysis may be applicable to other neuronal circuits such as the barrel cortex and other somatosensory regions as places where the anatomy may lend itself to following up such a connectivity analysis. However, our analysis is general and applies to any circuit where computation-specific ensembles arise and connectivity is approximately random, even if they lack the convenient anatomical demarcation of ensembles provided by the olfactory bulb or somatosensory regions.

Random connectivity in conjunction with strong dendritic nonlinearities, supports mixed selectivity of 3-4 different inputs

Mixed selectivity, wherein neurons are selective to combinations of task variables as opposed to single variables, has gained a lot of interest recently (Rigotti et al. 2013; Fusi, Miller, and Rigotti 2016; Eichenbaum 2018). Neurons showing mixed selectivity have been observed in the vibrissal cortex (Ranganathan et al. 2018), hippocampus (Stefanini et al. 2020), prefrontal cortex (Kobak et al. 2016) (based on (Romo et al. 1999; Brody et al. 2003)) among other brain regions. Non-linear mixed selectivity, wherein a neuron’s activity is not a simple weighted linear combination of the different stimuli, has the potential to expand the dimensionality of neural representations (Fusi, Miller, and Rigotti 2016). Such high dimensional representations are particularly helpful in tasks involving the mixing of multiple stimuli/variables because they enable simple downstream linear readout mechanisms to decode patterns.

How do neurons achieve mixed selectivity? A possible avenue is through dendritic nonlinearities. For example, work from (Ranganathan et al. 2018) in the vibrissal cortex has shown that mixed selectivity can arise from active dendritic integration in the layer 5 pyramidal neurons, which show joint representation for sensorimotor variables. This is particularly useful in the context of adaptive sensing. Our study explores this possibility of whether dendritic computation could act as a mechanism for neurons to achieve mixed selectivity. We show that redundancy in input representation in the form of large presynaptic ensembles yields postsynaptic neurons showing different kinds of selectivities: including pure selectivity, partially-mixed and fully-mixed selectivity. In fact, it is far more likely to find neurons showing partially-mixed selectivity than neurons showing pure/fully-mixed selectivity. While such partial mixing makes it difficult to decode the occurrence of full patterns, they may still act as building blocks for neurons further downstream to aid in the decoding process.

We show that under conditions of low noise, random connectivity coupled with strong dendritic nonlinearities provides an excellent substrate for neurons to mix different sources of information. Stimulus-driven groups of 3-5 inputs are likely with ensembles of 100 neurons (Figure 2G) in 4 of the 6 network configurations tested. This mixing occurs over short segments of dendrites, enabling each neuron to potentially mix multiple different combinations of 3-5 inputs. Since partially mixed groups are more likely than fully-mixed groups, decoding the identity of patterns would require further processing based on the population response of all such neurons.

Effective pattern representation through dendritic computation needs more than structural and functional connectivity

A key prediction of our study is that structural and even functional connectivity are necessary but not sufficient conditions to achieve postsynaptic computation in clusters and sequences. Several additional factors influence the effective selectivity for clustered/sequential computation. Our analysis of the hippocampal and cortical configurations occurring over three spatiotemporal scales (electrical, CICR and chemical) indicates that background activity plays a pivotal role in determining the ability of the network to extract useful information from activity patterns. Hence we found that the hippocampal configurations are better suited for grouped as well as sequential computations, as compared to the cortex (Figure 2, 4).

Our analysis sets boundary conditions on likely postsynaptic mechanisms for dendritic cluster computations. Specifically, chemical computations occurring over timescales of several seconds are vulnerable to background inputs whose likelihood of hitting a synapse scales with the duration of the computation. (Figure 2H, I). This effect is even more pronounced in the cortex due to its higher background activity (Buzsáki and Mizuseki 2014). Hence we posit that chemical timescales are less useful for ongoing activity-driven computation leading to changes in cell firing, and rather more useful from the context of plasticity. Chemical processes occurring over seconds to minutes, can modulate synaptic weights locally within dendritic clusters, strengthening grouped inputs through cooperative plasticity (Harvey et al. 2008). Similarly, our analysis showed that electrical mechanisms worked well for grouped computations, but they were quite sensitive to ectopic inputs in the case of sequences. Further, some studies suggest that 3-5 inputs may be insufficient to trigger dendritic spikes (Gasparini, Migliore, and Magee 2004), although (Goetz, Roth, and Häusser 2021) argue that a few very strong synapses can also trigger them. CICR operating on the timescales of ∼100-1000 ms, on the other hand, could be a potential mechanism that could influence both computation and plasticity. Higher level of dendritic Ca2+ affects activity at the soma through dendritic spikes (Palmer et al. 2014) and enhances dendrite-soma correlations (O’Hare et al. 2022). In addition, Ca2+ is known to trigger downstream mechanisms that can modulate plasticity (Mateos-Aparicio and Rodríguez-Moreno 2020). Hence, CICR is a promising candidate that could boost the impact of dendritic computations in a network context.

Sequence selectivity required low background activity and strong nonlinearities to amplify sequential inputs over other kinds of inputs at the dendritic level. Chemical mechanisms showed stronger selectivity than electrical mechanisms, and they were less prone to ectopic inputs arriving outside the zone of the sequence (Figure 5, 6). However, despite using strong nonlinearities at the dendritic zone, neurons receiving perfect sequential connectivity (PSCSD) showed weak selectivity at the somatic level when we accounted for inputs arriving across the entire dendritic arbor.. The sheer number of background or non-sequential ensemble inputs impinging on the dendritic arbor led to elevated baseline activation. This problem could be alleviated through precise EI balance. If the neurons exist in the state of precise EI balance (Hennequin, Agnes, and Vogels 2017; Bhatia, Moza, and Bhalla 2019), inhibition could cancel out the effect of background activity thus amplifying the contribution of sequential inputs to the neuron’s activation.

Thus, strong nonlinearities, timescales of a few 100 ms, and low noise regimes are best suited for both grouped and sequence computations. Precise EI balance in the background inputs improves the conversion of dendritic sequence selectivity to somatic sequence selectivity.

Overall, our study provides a general framework for assessing when dendritic grouped or sequence computation can discriminate between different patterns of ensemble activations in a randomly connected feedforward network. Our approach may need further detailing to account for area-specific features of connectivity, synaptic weights, neuronal morphologies, dendritic branching patterns and network architectures. Conversely, deviations from our theory may suggest where to look for additional factors that determine connectivity and neuronal selectivity. Our analysis touches lightly on the effects of inhibition and plasticity, though these too may contribute strongly to the outcome of clustered input. While connectivity is a necessary substrate for clustered dendritic computations, we show that factors such as background activity, length and timescales of postsynaptic mechanisms, and dendritic nonlinearities can play a key role in enabling the network to tap into such computations. In addition, our analysis can guide the design of biologically-inspired networks of neurons with dendrites for solving a variety of interesting problems (Chavlis and Poirazi 2024; Iyer et al. 2022; Asabuki and Fukai 2020; Asabuki, Kokate, and Fukai 2022).

Methods

The derivations for the equations have been included in the main text and in the appendix. All simulations used python version 3.8.10. We used both analytical equations and connectivity based simulations for the results shown in Figures 3 and 5. Simulations were run using a 128 core machine running Ubuntu server 22.04.2 LTS. All code used for this paper is available at github (https://github.com/BhallaLab/synaptic-convergence-paper). The readme file in the github repository contains instructions for running the simulations.

For simulations shown in Figures 2B and 5, the MOOSE Simulator (Ray and Bhalla 2008) version 4.0.0 was used. The NEURON simulator was used for Figure 6. Version numbers of all the python packages used are available in the requirements.txt file on the github repository mentioned above.

Connectivity based simulations

For both sequences and groups, we considered a feedforward network containing a presynaptic population of Tpre neurons. Connectivity to neurons in the postsynaptic population was simulated by sampling inputs randomly from the presynaptic population onto a single long dendrite of length L, which was discretized into p*Tpre(=20000) uniformly distributed synapses. Here p is the connection probability between the two layers. In each run connectivity is sampled for 4000 neurons. Hence the connectivity to the dendrite was represented as a matrix of 4000 neurons x 20000 inputs. As simulating connectivity and examining groups/sequences for the whole postsynaptic population in a serial manner was computationally time consuming, we parallelized the process by breaking down the simulation into 100 runs consisting of 4000 postsynaptic neurons per run. This does not affect the results since the connectivity to each neuron is sampled independently, and the presynaptic activity was kept the same across different runs by initializing the same seed for generating it. However, since we wanted neurons with different connectivities, the seed for generating connectivity was changed in each run. Hence we sampled connectivity for a total of 400,000 postsynaptic neurons in all simulations except for figure 7, where we sampled 1,000,000 neurons.

Generating presynaptic activity

The first N*M neurons in the presynaptic population represented the activity of M ensembles consisting of N neurons each. The remaining neurons participated in background activity, showing Poisson firing. A total of 100 trials, consisting of 3 trial types, (connectivity based, stimulus trials, and background only trials as defined below), were simulated. In the connectivity based trial, the ensemble participation probability, pe, is set to 1, and there is no background activity. A single trial of this type was run to estimate connectivity based groups/sequences. In stimulus trials, pe was set to 0.8 and the non-ensemble neurons are considered to be firing with a rate R in a Poisson manner. Background only trials had no ensemble activity. All neurons in background-only trials were considered to be participating in background activity with a firing rate R. With the Poisson assumption, the probability of a neuron firing in duration D due to background activity was considered to be pbg= 1 - e-RD. In each trial, the activity of each presynaptic neuron was sampled from a Bernoulli distribution with p=pe or p=pbg depending on whether it was an ensemble neuron or was participating in background activity. The activity of every neuron in the presynaptic (input) population was represented as a one, minus one, or zero, where ones specified that the neuron participated in ensemble activity, minus-ones represented background activity and zeros represented no activity.

Estimating the probability of occurrence of groups

In the case of groups, neurons in all M ensembles were coactive within duration D with a probability of pe=0.8 in stimulus trials and pe=1 in connectivity based trials. For each postsynaptic neuron an input array was generated where the entry at index i represented the ensemble ID of input arriving at position i on the dendrite if the input belonged to one of the ensembles. It was −1 if the input was due to background activity, and 0 if the synapse was inactive. The input array is scanned for the occurrences of groups by sliding a window of length Z, one synapse at time. For a group of size M, if there was at least one input from each of the M ensembles within zone Z, it was considered a fully-mixed group. A stimulus-driven group was defined as one that received M or more inputs in zone Z from neurons belonging to any of the ensembles. If there were M or more inputs from background activity in a window of length Z, it was considered a noise group. We defined any group as one consisting of M or more inputs, either from the ensembles or from background activity or a combination of the two, arriving in a zone of length Z. The sum of the number of groups occurring on the dendrite was calculated for each group type (true, stimulus-driven, noise and any). The probability of a group occurring on a neuron was estimated by calculating the mean number of neurons in the postsynaptic population that received at least one group of that specific type somewhere along the dendrite.

Estimating neuronal activation for the context of grouped computation

To explore the effect of nonlinear grouped dendritic computation on neuronal response, we applied a nonlinear function on local dendritic inputs. Using the input array generated in the previous step for detecting groups, the stretched exponential function depicted in the inset in Figure 3A was applied to the total number of inputs arriving on the dendrite within a zone of length Z. In this manner, using a rolling window approach the local dendritic activation at each location on the dendrite was estimated as the output of this nonlinearity. Finally, the net activation of the neuron was measured as the summed activation across the whole dendritic arbor of length L, normalized by the total number of synapses in the zone of length Z.

Neuronal activation

Where ω represents the nonlinear function, δ(x) = 1 if an input arrived at location x, else it is 0. L is the total dendritic length, Z the zone length and σ the inter-synapse spacing.

The nonlinearity adopted here was inspired from the cumulative distribution function of the Weibull distribution. It has the following useful properties:

  1. It is bounded, i.e. the local activation always saturates at 1.

  2. It is nonlinear.

  3. By controlling the exponent, one can control the effect fewer inputs have on the activation. This is particularly relevant for tuning the contribution smaller groups have on the activation of the neuron.

  4. One can also control the fixed value of this distribution, which is the point that remains frozen upon changing powers. We have used a value of 10 as the base of the second term, which implies that when all M inputs constituting a group are present the resulting local activation is 0.9, i.e. 90% of the maximum. When there are >M inputs, the activation saturates at a value of 1.

Estimating the probability of occurrence of sequences

A key difference between groups and sequences was that the presynaptic activity was simulated for a single time step in the case of groups. However, it was simulated over MAX_M(=9) time steps which was the maximum sequence length we considered. Hence the presynaptic activity for sequences was represented as a matrix of ones, minus ones and zeros of dimensions Tpre x 9, where ones represented ensemble activity, minus ones represented background activity and zeros represented no activity. In stimulus trials, at any given time step, only one ensemble was considered to be active. Neurons belonging to inactive ensembles in a particular time step were considered to participate in background activity in that time step. Although the presynaptic activity was simulated for a total of MAX_M time steps, while counting the number of sequences for a sequence of length M, only the activity from the first M time steps was used, as that was the length of the sequence we were interested in. Connectivity was simulated between the pre and postsynaptic neurons as described earlier.

Using this setup we looked for the occurrence of different kinds of sequences. Sequences consisted of inputs that occurred within a spacing of S to S+Δ from an input that arrived at a previous time step. Perfectly-ordered stimulus sequences (POSS) were made of ones alone. Noise sequences consisted of minus ones alone. Any sequences were formed from either ones or minus ones or a combination of both.

Since the pairwise relationships between inputs occurring in adjacent time-steps was sufficient to capture whether two inputs followed the S to S+Δ spacing, identifying synapses that followed this rule across adjacent time-steps aids in determining whether they participated in a sequence. Hence, we constructed distance matrices between synapses that followed the above spacing rule for adjacent time windows wherein the Xt->(t+1)[i,j] entry in the matrix was 1 if synapse j received input in the time window t+1, following the arrival of input at synapse i in time window t, provided synapse j occurred within a spacing of S to S+Δ of synapse i. The total number of sequences consisting of M inputs occurring between different pairs of synapses was obtained by multiplying all pairwise distance matrices and summing the entries in the product.

Total number of sequences that occurred on a neuron

Distance matrices were constructed separately for the cases of true, noise and all sequences for each neuron and the number of different sequences in each case was calculated using the above equation. The probability of a sequence occurring on a neuron was estimated by calculating the mean number of neurons in the postsynaptic population that received at least one sequence of that specific type somewhere along the dendrite.

Estimating neuronal selectivity for the context of sequence computation

In order to estimate neuronal selectivity for ordered sequences, we stimulated the same set of ensembles in the presynaptic population in different temporal orderings, and measured the activation of neurons in the postsynaptic population. An input matrix was constructed, wherein the M[i,j]th entry was 1 if synapse j received an input at time step i. The local activation Q, at location x on the dendrite at time t was modeled as:

Here the first term captures the decay of local activation in the absence of new input, i.e. γ<1 . The second term represents the baseline contribution of inputs, δ(t,x) = 1 if an input arrived at location x at time t, else it is 0. The third term represents the nonlinear contribution of sequential input. Prior activation Q within a region of Δ a distance S away from location x, is combined multiplicatively with input arriving at x at time t and is scaled nonlinearly using η. It is also passed ψ which keeps the sequential contribution within bounds.

For the analysis in Figure 7, the parameters for this formulation were chosen such that in the absence of any ectopic inputs, the local selectivity for the perfectly-ordered sequence was ∼0.8. This value is close to the value we obtained for chemical selectivity shown in Figure 5. All parameter values have been included in the supplementary.

The total activation of a neuron was obtained by integrating Q over all synapses across time to give the response Ap for each pattern p. Selectivity for the ordered sequence is measured as, Selectivity = (Aseq - Amean) / Amax

Where Aseq is the response to an ordered pattern, Amean was the mean of responses to all patterns, Amax was the maximum response obtained amongst the set of patterns tested.

Ca2+-Calmodulin simulation

With the help of MOOSE Simulator, we used the Ca2+-CaM system to model selectivity for grouped inputs. The model involves Ca2+ binding to Calmodulin in a series of steps resulting in activated Ca2+-Calmodulin (Figure 2a). Ca2+ stimulus was delivered in a spatially localized manner with synapses spaced at 2 μm, or in a dispersed manner, with inputs spaced at 10 μm on a 100 μm long dendrite which had a diameter of 5 μm. The dendrite was discretized into 1-dimensional voxels of length 0.5 μm. The mean concentration of CaM-Ca4 in the dendrite, wherein each CaM molecule is bound to 4 Ca2+ was measured over time for the clustered and dispersed input cases. The parameters of the model have been provided in the supplementary material.

Chemical simulation - Bistable-switch model

The bistable switch model was adapted from (Bhalla 2017), which was available on ModelDB at https://modeldb.science/227318. We modified the original model to include an expression for parsing ectopic inputs. The original set of parameters from (Bhalla 2017) were retained for the bistable-switch model. Simulations were performed using MOOSE. For Figure 5B, inputs were delivered either as an ordered pattern [0, 1, 2, 3, 4], or a scrambled pattern [4, 1, 0, 3, 2] with consecutive inputs spaced at a distance of 3 μm on a 100 μm long dendrite that had a diameter of 10 μm. The time interval between inputs was 2s. For Figures 6D and 6E, an additional set of 23 scrambled patterns obtained from different permutations of the perfectly-ordered sequence were also tested. Since permutations of a sequence of length M can yield a total of M! patterns, we sub-sampled from this set to select every M!/(M-4)! th pattern to get a total of 24 patterns for M>3. Within this set the first pattern was the ordered sequence itself. For the case of M=3, only the 6 available patterns (=3! permutations) were used. Response was measured as the total amount of molecule A present in the dendrite summed over time. Using these responses selectivity was calculated with the help of Equation 5.

In Figure 5D, the reference selectivity was estimated using the above formula for a set of patterns made of 5 inputs each. Further, in addition to the inputs constituting a sequence, an ectopic input was delivered. Different spatiotemporal combinations of the position and time of arrival of the ectopic input were tested. For Figure 5E, we repeated this exercise for patterns of different lengths either without any ectopic (reference), with a single ectopic or with two ectopic inputs. A smaller set of spatiotemporal pattern combinations was used in Figure 5E to keep the computational complexity within reasonable bounds. In addition, we restricted the analysis to ectopics occurring within the zone of the sequence.

Effect of ectopic inputs on electrical sequences

We chose the model published on ModelDB (https://modeldb.science/140828) for examining the effect of ectopic inputs on the selectivity of electrical sequences (Branco, Clark, and Häusser 2010). The base code from ModelDB was modified and run using the NEURON Simulator. The original channel kinetics parameters post the fix to NMDA_Mg_T initialization, and the dendrite used for testing were retained as is from the ModelDB github repository (https://github.com/ModelDBRepository/140828). We tested out the effects of ectopics using the passive version of the model.

For figure 6B, a sequence of 5 inputs was delivered as an inward, outward and scrambled sequence and the voltage response at the soma was measured. For figure 6C, inputs were delivered sequentially and using a number of different scrambled orderings. The voltage response at the soma was measured in each case. Selectivity was calculated using Equation 5, wherein A represented the peak excitatory postsynaptic potentials in this case. This exercise was repeated for different time intervals between successive inputs and for different sequence lengths to obtain the selectivity vs input velocity curves.

For figure 6D, E the base code was slightly modified to accommodate an ectopic input in addition to the different sequential/scrambled patterns that were being delivered. First, an ectopic input was added at different locations arriving at different times relative to the zone of the regular patterns, and selectivity is estimated at each position. The reference curve shows selectivity in the absence of the ectopic input.

This exercise was repeated for sequences of 3-9 inputs with a subset of spatio-temporal combinations of the ectopic input. We sampled a combination of 3 locations (start, middle and end) relative to the zone of the regular inputs, and 3 time points (start, middle and end) for the ectopic input. For sequences of >=6 inputs, 121 pattern combinations of the regular inputs were sampled since sampling all permutations of the sequence was computationally expensive.

Appendix

Refer to Table 2 in the main text for parameter definitions.

Analytical derivations for different types of groups

Probability of occurrence of stimulus-driven groups

Expectation number of connections from neurons in the M ensembles, on a zone of length Z on the target neuron

Probability of a zone of length Z receiving M or more connections from any of the neurons in the M ensembles

Probability of a zone of length Z that receives at least M connections from any of the neurons in the M ensembles occurring anywhere along the dendritic length of a neuron,

Where kcSDG lies between and depending on the degree of overlap between zones in different network configurations.

Probability of a neuron receiving an active stimulus-driven group where there are at least M synapses receiving active inputs from neurons in the M ensembles is give by

Where

Where kcSDG again lies between and depending on the degree of overlap between zones.

Probability of occurrence of noise groups

Consider a network where the rate of background activity at synapses is R Hz.

Expected number of inputs arriving in duration D due to background activity at a single synapse = RD

Assuming that background activity in the presynaptic is poisson in nature,probability of a synapse being hit by background activity in activity in duration

Expected number of converging axonal connections in zone Z from the M ensembles

Expected number of synapses being hit by noisy inputs in a dendritic zone of length Z in duration

Where σ ≈ inter-synapse interval

Probability of inputs from background activity arriving at M or more synapses in a zone of length

Probability of a zone of length Z that has at least M synapses receiving background activity in duration

Where kNG lies between and based on the degree of overlap between zones.

Probability of occurrence of any group

Expected number of synapses being hit by noisy inputs in a dendritic zone of length Z in duration

Expected number of synapses in zone Z being hit by inputs from M ensembles

Expected number of synapses being hit either by ensemble inputs (from the M ensembles) or noisy inputs in a dendritic zone of length Z in duration

Probability of either ensemble inputs or background activity or a combination of them arriving at M or more synapses in a zone of length

Probability of a zone of length Z that has at least M synapses receiving any activity in duration

Where kNG lies between and depending on the degree of overlap between zones.

Analytical derivations for different types of sequences

Probability of occurrence of noise sequences

Assuming that background activity in the presynaptic population is Poisson in nature, probability of a synapse being hit by background activity in duration

Expected number of synapses being hit by background activity from any neuron in the first time step D, anywhere on target neuron

Expected number of inputs from background activity arriving in the second time-step within a spacing of S to S+Δ from an input that arrived in the first time step

Expectation number of sequences of M inputs coming from background activity in successive time steps, with the subsequent input arriving within a spacing of S to S+Δ with respect to the preceding input

Using Poisson approximation, probability of one or more background activity driven sequences of M inputs, occurring anywhere along the dendritic length of a neuron

Probability of occurrence of any sequence

Expected number of synapses being hit by either background activity or from activity in neurons in the first ensemble in the first time step D, anywhere on target neuron

Expected number of synapses being hit by either background activity or ensemble activity in the second time-step within a spacing of S to S+Δ from an input that arrived in the first time step

Expected number of sequences of M inputs coming from either background activity or ensemble activity in successive time steps, with the subsequent input arriving within a spacing of S to S+Δ with respect to the preceding input

Using Poisson approximation, probability of one or more sequences of M inputs that are driven by either background inputs or ensemble inputs or a combination of them, occurring anywhere along the dendritic length of a neuron

Probability of occurrence of gap-fill sequences

Gap-fill sequences are those that are built from a combination of ensemble and background inputs. Hence they contain at least one input from each of the two classes.

Expectation number of M-length gap-fill sequences constructed from inputs arriving in subsequent time windows of duration D, following the S to S+Δ spacing rule, on a neuron

Probability of occurrence of one or more gap-fill sequences containing M inputs anywhere on a neuron

Author contributions

Bhanu Priya Somashekar

Conceptualization, Formal analysis, Validation, Investigation, Visualization, Methodology, Writing—original draft, Writing—review and editing

Upinder Singh Bhalla

Conceptualization, Supervision, Funding acquisition, Investigation, Writing—original draft, Methodology, Project administration, Writing—review and editing

Acknowledgements

BPS and USB are at NCBS-TIFR which receives the support of the Department of Atomic Energy, Government of India, under Project Identification No. RTI 4006. We would like to acknowledge NCBS Supercomputing Facility for access to the cluster. We would like to specially thank Mukund Thattai for ideas and discussions on the project. In addition, we acknowledge Anal Kumar, Sahil Moza, Shaurya Rahul Narlanka, Sriram Narayanan, Sulu Mohan and Vinu Varghese Pulikkottil for valuable suggestions on the manuscript.

Supplementary information

Tables

- Molecular species used in the Ca2+-CaM model

- Reaction parameters of the Ca2+-CaM model

- Kappa values for different kinds of groups