Abstract
Understanding how different neuronal types connect and communicate is critical to interpreting brain function and behavior. However, it has remained a formidable challenge to decipher the genetic underpinnings that dictate the specific connections formed between pre- and post-synaptic neuronal types. To address this, we propose a novel bilinear modeling approach that leverages the architecture similar to that of recommendation systems. Our model transforms the gene expressions of mouse bipolar cells (presynaptic) and retinal ganglion cells (postsynaptic), obtained from single-cell transcriptomics, into a covariance matrix. The objective is to construct this covariance matrix that closely mirrors a connectivity matrix, derived from connectomic data, reflecting the known anatomical connections between these neuronal types. Our model successfully recaptiulates recognized connectivity motifs and provides interpretable insights into genetic interactions that shape the connectivity. Specifically, it identifies unique genetic signatures associated with different connectivity motifs, including genes important to cell-cell adhesion and synapse formation, highlighting their role in orchestrating specific synaptic connections between these neurons. Our work establishes an innovative computational strategy for decoding the genetic programming of neuronal type connectivity. It not only sets a new benchmark for single-cell transcriptomic analysis of synaptic connections but also paves the way for mechanistic studies of neural circuit assembly and genetic manipulation of circuit wiring.
1 Introduction
One of the fundamental objectives in neuroscience is understanding how diverse neuronal cell types establish connections to form functional circuits. This understanding serves as a cornerstone for decoding how the nervous system processes information and coordinates responses to stimuli [1]. Despite this, the genetic mechanisms determining the specific connections between distinct neuronal types, especially within complex brain structures, remains elusive [2, 3].
Recent advances in transcriptomics and connectomics provide opportunities to probe this. Single-cell transcriptomics enables high-resolution profiling of gene expressions across neuronal types [4, 5], while connectomic data offers detailed maps quantifying connections between neuronal cell types [6, 7, 8]. However, the challenge of linking gene expressions derived from single-cell transcriptomics to precise neuronal connectivity patterns evident from connectomic data has yet to be fully addressed.
Drawing inspiration from the field of machine learning, particularly recommendation systems, we introduce a bilinear model to bridge this gap. This model, in the context of recommendation systems, has been successful in capturing intricate user-item interactions [9]. By treating the gene expressions of pre- and post-synaptic neurons and their connectivity akin to users, items, and their ratings, we adapt the architecture of recommendation systems to the neurobiological domain. We hypothesize that a similar model could capture the complex relationships between genetic patterns of presynaptic and postsynaptic neurons and their connectivity.
Applying this model to single-cell transcriptomic and connectomic data from mouse retinal neurons, we demonstrate that it can effectively learn connectivity patterns between bipolar cells (BCs, presynaptic) and retinal ganglion cells (RGCs, postsynaptic). The model not only unveils connectivity motifs between BCs and RGCs but also provides biologically meaningful insights into candidate genes and the genetic interactions that orchestrate this connectivity. Furthermore, our model predicts potential BC partners for RGC transcriptomic types, with these predictions aligned substantially with functional descriptions of these cell types from previous studies. Collectively, this work significantly contributes to the ongoing exploration of the genetic code underlying neuronal connectivity and suggests a potential paradigm shift in the analysis of single-cell transcriptomic data in neuroscience.
2 Background: Synaptic Specificity between Neuronal Types
The intricate neural networks that form the basis of our nervous system are a product of specific synaptic connections between different types of neurons. This specificity is not a mere coincidence but a meticulously orchestrated process that underpins the functionality of the entire network [3]. Each neuron can form thousands of connections, or synapses, with other neurons, and the specificity of these connections determines the neuron’s function and, by extension, the network’s function as a whole. A classic example of this is seen in the retina, where different types of BCs form specific synaptic connections with various types of RGCs [7, 10, 11]. These connections create parallel pathways that transform visual signals from photoreceptors to RGCs, which subsequently transmit the information to the brain [12, 13].
The genetic principles guiding the formation of these specific connections, particularly in complex brain structures, remains elusive. The brain’s complexity, with its billions of neurons and trillions of synapses, poses significant challenges in identifying the specific genes and genetic mechanisms that guide the formation of these connections. Despite advances in genetic and neurobiological research, such as understanding the roles of certain recognition molecules and adhesion molecules in synaptic specificity, the genetic foundation of connectivity between neuronal types is still largely unknown [14, 3, 15].
Emerging tools and technologies offer unprecedented opportunities to unravel these mysteries. Among these, the transcriptome and connectome are particularly promising [3, 16]. The transcriptome, the complete set of RNA transcripts produced by the genome, can provide valuable insights into the genes that are active in different types of neurons and at different stages of neuronal development. This can help identify candidate genes that may play a role in guiding neuronal connectivity. The connectome, on the other hand, provides a detailed map of the connections between neurons. By combining information from the transcriptome and connectome, it is possible to link specific genes to specific connections, thereby shedding light on the genetic basis of neuronal connectivity.
3 Related Work: Collaborative Filtering
Our strategy draws inspiration from the concept of collaborative filtering using bilinear models, a technique fundamental to recommendation systems [17, 18]. These systems predict a user’s preference for an item (e.g., a movie or product) based on user-item interaction data.
Bilinear models capture the interaction between users and items via low-dimensional latent features [9, 19]. Mathematically, for user i and item j, we denote their original features as xi ∈ R1×p and yj ∈ R1×q, respectively. These features are then projected into a shared latent space with dimension d via transformations xiA (where A ∈ Rp×d) and yjB (where B ∈ Rq×d). The predicted rating of the user for the item is then formulated as:
In the context of collaborative filtering, the goal is to optimize the transformation matrices A and B to align the predicted rating rij with the ground-truth zij. This is expressed as the following optimization problem:
Or in the matrix form:
Here, the objective is to minimize the Frobenius norm of the residual matrix Z − (XA)(Y B)T.
In our study, we interpret neuronal connectivity through the lens of recommendation systems, viewing presynaptic neurons as “users”, postsynaptic neurons as “items”, and the synapses formed between them as “ratings”. Our chosen bilinear model extracts latent features of pre- and post-synaptic neurons from their respective gene expressions. One key advantage of the bilinear model is its capacity to assign different weights to the gene expressions of pre- and post-synaptic neurons, enabling the model to capture not just homogeneous but also complex, heterogeneous interactions fundamental to understanding neuronal connectivity. Prior studies have highlighted such heterogeneous interactions, noting the formation of connections between pre- and post-synaptic neurons expressing different cadherins, indicative of a heterogeneous adhesion process [20, 21].
4 Bilinear Model for Neuronal Type Connectivity
4.1 Objective Functions
We discuss the bilinear model for neuronal type connectivity in the following two scenarios: the first in which gene expression and connectivity of each cell are known simultaneously and the second where connectivity and gene expressions of neuronal types are from different sources. The bilinear models for these two situations are illustrated in Figure 1.
4.1.1 Gene Expression and Connectivity of Each Cell are Known Simultaneously
We begin with an ideal scenario where both the gene expression profiles and connectivity of individual cells are known concurrently. In this setting, we have a presynaptic neuronal types and b postsynaptic neuronal types, indexed by i and j, respectively. Each type contains a number of neurons, signified as ni for presynaptic and nj for postsynaptic types. The gene expression vector for the kth cell in the presynaptic type i is designated as x(ik), where k ∈ 1, 2, …, ni, while for the lth cell in postsynaptic type j, it is y(jl) with l ∈ 1, 2, …, nj. We depict the connectivity metric between a presynaptic neuron and a postsynaptic neuron as z(ik)(jl).
Drawing from the principles of collaborative filtering, we develop the following optimization objective:
Here, A and B denote the transformation matrices we aim to learn. This formula can also be expressed in its matrix form as:
In this equation, W symbolizes a weight matrix where each element . As our study focuses on the genetic code of pre-and post-synaptic neuronal types rather than individual neurons, this weight matrix ensures that the model does not disproportionately favor neuronal types with a greater number of neurons over rarer types.
In the context of high dimensionality of gene expressions, the bilinear model may face a common issue in machine learning called multicollinearity, a condition where one or more predictor variables are highly correlated. To overcome this, we can perform principal component analysis (PCA) on the gene expression vectors, transforming them into a new coordinate system. By excluding components whose negligible eigenvalues, we effectively get rid of redundant information, thus mitigating the effects of multicollinearity. This transformed gene expression data can then be used in our bilinear model, leading to more stable and reliable estimates of connectivity. In mathematical terms, applying PCA to the zero-centered and unit-variance adjusted matrices X and Y results in the approximations:
and
where U and V indicate the PCA transformation of X and Y respectively. The original optimization problem now becomes:
where X′ = XU, Y ′ = Y V, A′ = U T A, B′ = V T B. It’s worth noting that if the optimization problem is solved in the PCA-transformed space, we need to transform A′ and B′ back A and B in order to retrieve the corresponding weight of each gene.
4.1.2 Connectivity and Gene Expressions of Neuronal Types are from Different Sources
In real scenarios, gene expression profiles and connectivity information are often derived from separate sources, such as single-cell sequencing [22, 23] and connectome data [7, 24, 25]. Bridging these datasets requires classifying neurons into cell types based on their gene expression profiles and morphological characteristics. These cell types from different sources are subsequently aligned according to established biological knowledge (e.g., specific gene markers are known to be expressed in certain morphologically-defined cell types [26]).
The primary challenge in this scenario is that, while we can align cell types (denoted by indices i and j in equation 4), we are unable to associate individual cells (represented by indices k and l in equation 4). To tackle this issue, we adopt a simplifying assumption that the connectivity and latent gene expression features of individual cells can be approximated by the averages of their corresponding cell types. This premise hinges on the notion that the connectivity metrics and latent gene expression features of individual cells are close enough to the mean value of their corresponding cell types.
As a result, our optimization objective in equation 4 becomes:
In this equation, z(i.)(j.) denotes the mean connectivity metric between presynaptic cell type i and postsynaptic cell type j. Meanwhile, x(i.) and y(j.) represent the average gene expressions of cell types i and j respectively.
While optimizing the transformation matrices A and B, we impose constraints on these matrices to ensure that the variance of latent gene expression features within each neuronal type is minimized. Specifically, we define ϵ as a small enough value and impose the following constraints on A:
where
and B:
where
These conditions assure that the latent gene expression features of individual cells are proximate enough to the average value within their respective cell types. With these constraints in mind, we formulate the optimization problem as follows:
In this equation, denotes the average gene expressions of the a presynaptic cell types, wherein each element is indicative of the average gene expression feature m within cell type i. Likewise, represents the average gene expressions of the b postsynaptic cell types, with each element signifying the average gene expression feature m in cell type j.
In practical application, we approximate Σx and Σy with their diagonal estimates and [27, 28]. We then transform the initial optimization problem into the following:
Here, elements in are defined as and elements in Ŷ ∈ Rb × q are given by . The optimization of this formulation tends to be compuqationally more tractable.
In summary, our methodology adapts when gene expression profiles and the connectivity matrix originate from distinct sources. Instead of aligning at the level of individual cells, we focus on the alignment of neuronal types. We achieve this by mapping gene expressions into a latent space via transformation matrices  and , with the optimization process aiming to minimize the discrepancies between these two sources of information while maintaining consistency of the gene expression features within individual neuronal types.
4.2 Optimization Algorithm
To solve the optimization problem as outlined in equation 15, we construct the following loss function:
This function can be regarded as the Lagrangian of the initial problem. Here, λA and λB are treated as hyperparameters, and their optimal values are determined through a grid search.
Given this loss function, we propose an alternative gradient descent algorithm to find the solutions. This algorithm alternates between updating the transformation matrices  and , using the gradient descent optimization method.
The algorithm begins by initializing transformation matrices  and using random values drawn from a standard normal distribution. The central aspect of the algorithm is an iterative loop that alternates the updates of transformation matrices  and .
Algorithm 1
Alternative Gradient Descent (AGD) for Bilinear Mapping
During each iteration, the algorithm computes the predicted connectivity metric using the current estimates of , and Ŷ. Subsequently, the gradient of the loss function with respect to the transformation matrices is calculated, and the matrices are updated by moving in the negative gradient’s direction.
This iterative process is repeated until the transformation matrices  and converge to a steady solution. Upon completion, the algorithm yields the optimized transformation matrices.
This gradient descent-based algorithm provides a computationally efficient solution to the bilinear mapping problem between gene expression profiles and connectivity metrics while adhering to the constraints unique to our problem design. As a result, it produces associations between gene expression profiles of cell types and their connectivity.
5 Datasets and Pre-processing
Our study hinges on two distinct data collections of mouse retina neurons: single-cell transcriptomic data from previous studies and connectomic data from the EyeWire project. Together, these datasets provide us with connectivity information and gene expression profiles, which constitute the key ingredients for our proposed bilinear model.
5.1 Single-cell Transcriptomic Data
The single-cell transcriptomic data used in our study include the gene expression profiles for two classes of mouse retina neurons - presynaptic BCs as reported by Shekhar et al. [22], and postsynaptic RGCs as reported by Tran et al. [23].
Preprocessing of this data adhered to previously documented procedures [22, 23, 29]. The transcript counts within each cell were first normalized to align with the median number of the transcripts per cell, followed by a log-transformation of the normalized counts. High variable genes (HVGs) were then selected using an approach based on establishing a relationship between mean expression level and the coefficient of variance [30, 31, 32]. We focused on those cells whose types correspond with the neuronal types outlined in the connectomic data, as delineated later in Table 1, Table 2, and Table 3. This yielded two matrices, X and Y, representing presynaptic BCs and postsynaptic RGCs, where each row pertains to a cell and each column represents an HVG. The dimensions of X and Y are 22453 × 17144 and 3779 × 12926, respectively.
Next, we performed a principal component analysis (PCA) on these matrices to transform the gene expression data into the principal component (PC) space. We retained only the PCs that account for a cumulative 95% of explained variance. Consequently, the gene expression of the BCs in X and the RGCs in Y were featurized by their respective PCs, resulting in matrices of dimensions 22453 × 11323 and 3779 × 3142, respectively.
Based on each cell’s neuronal type, we computed the variance of gene expression features within these types. Mathematically, the variance of gene expression feature m within the BC types and the RGC types are expressed as:
Taking and to represent the average gene expression feature m of the BC type i and the RGC type j, we were able construct matrices, and Ŷ, in which and . In these matrics, each row represents a cell type, with the dimensions of being 25×11323 and Ŷ being 12×3142. These matrices serve to bridge the gene expression of BC types and RGC types with the connectivity matrix of these neuronal types derived from the connectomic data.
5.2 Connectivity Data
The connectivity matrix of neuronal types is derived from connectomic data acquired through the process of serial electron microscopy (EM)-based reconstruction of brain tissues [6, 7, 8]. From these reconstructed tissues, connectivity measurements are usually expressed as either the contact area or the number of synapses between neurons [7, 33]. When normalized to the total contact area or total number of synapses of each neuron, the resulting metric, ranging from 0 to 1, signifies the percentage of contact area or synapses formed between neurons. This normalized metric provides a quantitative connectivity measure, where 0 indicates no connectivity and 1 implies complete connectivity between two neurons.
Our analysis utilized the neural reconstruction data of mouse retinal neurons, courtesy of the EyeWire project, a crowd-sourced initiative that generates 3D reconstructions of neurons from serial section EM images [34]. This extensive dataset facilitated the derivation of a comprehensive connectivity matrix between two classes of mouse retina neurons - BCs [25] and RGCs [24]. The data were sourced from the EyeWire Museum (https://museum.eyewire.org/), which offers detailed information for each cell in a JSON file, including attributes like “cell id”, “cell type”, “cell class”, and “stratification”. The stratification profile describes the linear density of voxel volume as a function of the inner plexiform layer (IPL) depth [34, 25, 24].
We approximated the connectivity metric between a BC and a RGC using the cosine similarity of their stratification profiles. Let vik and vjl denote the stratification profiles of the kth cell in BC type i and the lth cell in RGC type j, respectively. The connectivity metric z(ik)(jl) between these two neurons can be expressed as:
This equation represents the degree of overlap in their voxel volume profile within the IPL, resulting in the connectivity matrix Z between mouse BCs and RGCs. To allow for both positive and negative values within the matrix, we standardized by subtracting the mean of and then dividing by its standard deviation. Subsequently, the connectivity matrix between mouse BC and RGC neuronal types was calculated, with each element representing the average of the connectivity metrics between cells of BC type i and cells of RGC type j.
5.3 Correspondence of Cell Types between Datasets
Aligning neuronal types as annotated in the single-cell transcriptomic data and those identified in the connectomic data was informed by findings from previous studies. Notably, a one-to-one correspondence exists between BC cell types classified by Shekhar et al. [22] and Greene et al. [25]. This correspondence is presented in Table 1.
Regarding RGC types, alignment between cell types annotated in Tran et al. [23] and Bae et al. [24] was established primarily based on the findings from Goetz et al. [26]. This study presents a unified classification of mouse RGC types, based on their functional, morphological, and gene expression features. The corresponding RGC types were mainly obtained from Supplementary Table S3 of Goetz et al. (Table 2): with the following additions derived from Supplementary Table S1 of Tran et al., based on the expressions of genetic markers of the following RGC types (Table 3):
We carried out subsequent analyses based on the alignment of these neuronal types.
6 Model Training and Validation
With the bilinear model outlined in Section 4, we iteratively optimized transformation matrics, denoted as  and , using the ADG optimization algorithm. The goal was to minimize the prescribed loss function. The initial states of these transformation matrices were randomly generated from a standard normal distribution, and updates were continued until either the change in loss fell below a predetermined threshold of 10−6 or a maximum iteration count of 106 was reached.
We examined two sets of hyperparameters during the optimization process: the regularization parameters, λA and λB, and the dimensionality of the latent feature space. Preliminary tests suggested that a lower loss was achieved when λA and λB were set to equivalent values. Consequently, both were unified under a single parameter, λ.
We utilized 5-fold cross-validation to identify the optimal hyperparameters. This procedure partitioned the connectivity matrix entries into five unique subsets or “folds”. The model was then trained using a combination of four folds, with the remaining one serving as a validation set. This procedure was repeated five times, with each iteration reserving a different fold for validation. Throughout this cross-validation process, we varied λ across the range [0.1, 1, 10, 100] and the latent feature space’s dimensionality across the range [1, 2, 3, 4, 8].
Figure 2a presents a heatmap of the logarithm (base 10) of the validation loss, highlighting variations of λ and dimensionality. It can be observed that the lowest validation loss is achieved when λ equals 1 and the dimensionality equals 2, as shown in Figure 2b,c. These specific values were thus chosen as the optimal hyperparameters. With these optimal hyperparameters, we performed the final round of training on the entire dataset to yield the definitive transformation matrices, Â and , using a learning rate of 5 × 10−7.
7 Results
7.1 Bilinear Model Reconstructs Neuronal Type-Specific Connectivity Map from Gene Expression Profiles
Upon completion of the final training process, our optimized bilinear model produced transformation matrices, Â and . We used these matrices to project the normalized single-cell transcriptomic data,
and Ŷ, into a shared latent feature space. Consequently, we obtained projected representations for BC and RGC types, and , respectively. With these latent representations, we were able to reconstruct the cell-type-specific connectivity matrix: (Figure 3a).
To evaluate our model, we compared the reconstructed connectivity matrix with the one derived from connectomic data (Figure 3b). We calculated the Pearson correlation coefficient between entries of the two matrices to assess their agreement. The resulting correlation of 0.83 (p < 0.001) demonstrated a robust association between the transformed gene expression features and the connectomic data. This result attests to our model’s capability in capturing the relationship between these two distinct types of biological information.
7.2 Bilinear Model Recapitulates Recognized Connectivity Motifs
Our cross-validation procedure indicated that the optimal number of latent dimensions was two. This finding suggested that these two dimensions capture the essential connectivity motifs between BC and RGC types. This led us to further investigate what are these motifs and how they are different from each other.
We first reconstructed connectivity using only the first latent dimension. The first dimension appeared to emphasize connectivity patterns between BCs and RGCs that laminate within the IPL’s central region, as well as those that laminate within the marginal region (Figure 4a,d,g). We then reconstructed connectivity using only the second latent dimension. Notably, the spotlight shifted to connections between BCs and RGCs that laminate within the outer and inner regions of the IPL, respectively (Figure 4b,e,i).
To confirm these observations, we further visualized BC and RGC types within the two-dimensional latent feature space (Figure 4c,f). Grouping BC and RGC types based on whether they fell within the positive or negative halves of the latent dimensions, we color-coded their stratification profiles within the IPL by group. BCs and RGCs that fell within the positive half of latent dimension 1 tend to stratify within the IPL’s central region, delineated by the boundaries formed by the ON and OFF starburst amacrine cells (SACs) (Figure 4d,g). Conversely, those falling within the negative half of this dimension tend to stratify in the marginal region of the IPL. As for the second latent dimension, BCs and RGCs that fell within the positive half predominantly stratify in the inner region of the IPL (Figure 4e,i), while those within the negative half primarily stratify in the IPL’s outer region.
Interestingly, these distinct connectivity motifs align with two widely recognized properties of retinal neurons: kinetic attributes that reflect the temporal dynamics (transient versus sustained responses) of a neuron responding to visual stimuli, and polarity (ON versus OFF responses) reflecting whether a neuron responds to the initiation or cessation of a stimulus [35, 10, 11, 36]. This correlation implies that our bilinear model has successfully captured key aspects of retinal circuitry from gene expression data.
7.3 Bilinear Model Reveals Interpretable Insights into Gene Signatures Associated with Different Connectivity Motifs
The inherent linearity of our bilinear model affords a significant advantage: it enables the direct interpretation of gene expressions by examining their associated weights in the model. These weights signify the importance of each gene in determining the connectivity motifs between the BC and RGC types. We identified the top 50 genes with the largest positive or negative weights for BCs and RGCs across both latent dimensions. We plotted their weights alongside their expression profiles in the respective cell types (Figure 5a-d).
Our analysis unveiled distinct gene signatures associated with the connectivity motifs revealed by the two latent dimensions. In the first latent dimension, genes like CDH11 and EPHA3, involved in cell adhesion and axon guidance, carried high weights for BCs forming synapses in the IPL’s central region. In contrast, for BCs synapsing in the marginal region, we observed high weights in the cell adhesion molecule PCDH9 and the axon guidance cue UNC5D (Figure 5a). This pattern was echoed in RGCs but involved a slightly different set of molecules. For example, in RGCs forming synapses in the IPL’s central region, the cell adhesion molecule PCDH7 carried high weights, whereas for RGCs synapsing in the marginal region, cell adhesion molecules PCDH11X and CDH12 were associated with high weights (Figure 5b).
The second latent dimension revealed a comparable pattern, albeit with different gene signatures. For BCs laminating in the IPL’s outer region, high weights were assigned with guidance cues such as SLIT2, NLGN1, EPHA3 and PLXNA4, as well as the adhesion molecule DSCAM. For BCs in the inner region, the adhesion molecule CNTN5 was associated with a high weight (Figure 5c). In RGCs, we noticed that guidance molecules such as PLXNA2, SLITRK6 and PLXNA4 along with adhesion modules CDH8 and LRRC4C were associated with high weights for cells forming synapses in the IPL’s outer region. In contrast, the adhesion molecule SDK2 was among the top genes for RGCs laminating and forming synapses in the IPL’s inner region (Figure 5d). Some of these genes or gene families, such as Plexins (PLXNA2, PLXNA4), Contactin5 (CNTN5), Sidekick2 (SDK2), and Cadherins (CDH8,11,12), are known to play crucial roles in establishing specific synaptic connections [37, 38, 39, 40, 20, 21, 41]. Others, particularly delta-protocadherins (PCDH7,9,11x), emerged as new candidates potentially mediating specific synaptic connections [3].
To elucidate the biological implications of these identified gene sets, we further conducted Gene Ontology (GO) enrichment analysis on the top genes through g:Profiler, a public web server for GO enrichment analysis [42, 43]. This tool allowed us to delve into the molecular functions, cellular pathways, and biological processes associated with these genes. Intriguingly, when we plotted the top 10 significant GO terms for BCs and RGCs in latent dimension 1 or 2 (Figure 5e-h), we found two common themes associated with the top genes: neuronal development and synaptic organization. This observation underlines the potential role of the top genes in forming and shaping the specific connections between BC and RGC types.
7.4 Bilinear Model Predicts Connectivity Partners of Transcriptomically-Defined RGC Types
The success of recommendation systems in accurately predicting the preferences of new users inspired us to leverage the bilinear model for predicting the connectivity partners of RGC types whose interconnections with BC types remain uncharted. There are some RGC types defined from single-cell transcriptomic data [23], which lack clear correspondence with those identified through connectomics studies [24]. This discrepancy leaves the connectivity patterns of these transcriptionally-defined RGC types unknown, providing an opportunity for our model to predict their BC partners.
To accomplish this, we first projected these RGC types into the same latent space as those used to train the model (Figure 6a). We then employed this projection to construct a connectivity matrix between these RGC types and BC types (Figure 6b), facilitating educated estimates about their connectivity partners. For each transcriptionally-defined RGC type, we identified the top three BC types as potential partners, determined by the highest values present in the connectivity matrix. These three BC types could provide insight into the potential synaptic input to each RGC type. Detailed predictions are presented in Table 4.
Although the ground truth connectivity of these RGC types remains unknown due to the absence of matching types in connectomic data, Goetz et al. [26], via Patch-seq, attempted to match some transcriptomic types with functionally defined RGC types. These functional descriptions may hint at the BC partners of these RGC types. For instance, an RGC exhibiting OFF sustained responses is likely to be synaptically linked with BC types bc1-2, known to mediate OFF sustained pathways. Conversely, an RGC that displays ON sustained responses likely receives synaptic inputs from BC types bc6-9, which oversee ON sustained pathways. We summarized these functional descriptions in Table 4, referencing Figure 5A from Goetz et al. [26], and highlighted whether our predictions were consistent with these functional annotations. Among the ten predictions made, eight aligned with these functional descriptions, lending support to the predictive power of our model.
8 Discussion
This study showcased an innovative computational strategy that integrates transcriptomic and connectomic data to elucidate genetic principles dictating neural circuit wiring. Leveraging a bilinear model, we reconstructed a neuronal type-specific connectivity map from gene expression profiles. We have demonstrated the model’s capability to recapitulate recognized connectivity motifs of the retinal circuit and provide interpretable insights into the gene signatures associated with these motifs. Moreover, our model can predict potential connectivity partners for transcriptomically-defined neuronal types whose connectomes are yet to be fully characterized. These make our bilinear model a valuable tool for investigating gene-regulatory mechanisms involved in neural circuit wiring.
The inspiration for our bilinear model stemmed from recommendation systems – a machine learning domain focused on capturing intricate interactions between users and items and predicting user preferences in commercial settings. This analogy served as a useful framework in our study, where the roles of users and items in the recommendation systems were mirrored by presynaptic and postsynaptic neurons, respectively. Likewise, the user-item preference matrix corresponds to the synaptic connection matrix in neural circuits. The recommendation systems are based on the assumption that user preferences and item attributes can be represented by latent factors; similarly, our model assumes that synaptic connectivity between various neuron types is determined by a shared latent feature space derived from gene expression profiles.
Our bililinear model successfully recapitulated two core connectivity motifs of the retinal circuit, representing synapses formed in central or marginal parts of the IPL, and synapses formed in outer or inner regions. These motifs align well with recognized properties of retinal neurons: kinetic attributes (transient versus sustained responses) and polarity (ON versus OFF responses). Significantly, these motifs aren’t predefined or explicitly encoded into the model; instead, they emerge naturally from the model, further attesting to the model’s power to capture key aspects of retinal circuitry.
The bilinear model also revealed unique insights into the gene signatures associated with the connectivity motifs. The weight vectors in the transformation matrices provide a means to assess the relative importance of individual genes. This direct interpretability is a significant advantage of the linear model, allowing for a more intuitive understanding of the gene-to-connectivity transformation process. Our analysis discovered distinct gene signatures associated with different connectivity motifs. Among these genes, some have been previously implicated in mediating specific synaptic connections, thererby validating our approach. For instance, Plexins A4 and A2 (PLXNA4, PLXNA2), predicted to be crucial for RGCs’ synapsing in the outer IPL, have been shown to be necessary for forming specific lamina of the IPL in the mouse retina, interacting with the guidance molecule Semaphorin 6A (Sem6A) [37, 38]. Contactin5 (CNTN5), which our model predicted as vital for BCs forming synapses in the inner IPL, has been shown to be essential for synapses between ON BCs and the ON lamina of ON-OFF direction-selective ganglion cells (ooDSGCs) [39]. Sidekick2 (SDK2), predicted to be critical for RGCs’ synapses in the inner IPL, has been shown to guide the formation of a retinal circuit that detects differential motion [40]. Similarly, Cadherins (CDH8,11,12), whose combinations have been implicated in synaptic specificity within retinal circuits [20, 21], were highlighted for multiple connectivity motifs. In particular, Cadherin8 (CDH8), which our model predicted to be crucial for RGC’s synaptic connections in the outer IPL, has been shown to be guided by the transciptional factor Tbr1 for laminar patterning of J-RGCs, a type of OFF direction-selective RGCs [41].
In additional to these validated gene signatures, our analysis identified promising candidate genes that may mediate specific synaptic connections. Particularly, delta-protocadherins (PCDH7,9,11x) appeared as potential new candidates. While their roles in synaptic connectivity aren’t fully understood [3], mutations in delta-protocadherins in mice and humans have been linked with various neurological phenotypes, including axon growth and guidance impairments and changes in synaptic plasticity and stability [44, 45, 46]. Future experimental studies are needed to validate these findings and further unravel the roles of these genes in neural circuit formation and function.
While our approach was illustrated using the mouse retina, it holds vast potential to decipher the genetic programming of neuronal connectivity in other areas of the nervous system, given sufficient data on both gene expression profiles and synaptic connections. For instance, single-cell sequencing across multiple cortex regions and the hippocampus in mice has defined transcriptomic cell types [47, 48, 49]. Concurrently, considerable progress has been made in connectomic studies of mouse cortices, particularly the visual cortex [50, 51, 33, 52]. Integrating these data sources enables a comprehensive examination of the genetic underpinnings of connectivities in these brain regions. Our approach serves as a pioneering step in investigating how gene expression patterns contribute to the diversity of neuronal circuits across different brain regions, thereby setting the stage for a holistic understanding of the genetic blueprint governing neuronal circuit wiring throughout the entire brain.
9 Future Directions
This study introduces a novel and powerful methodology for inferring connectivity maps from gene expression data, bridging the gap between gene expression and synaptic connectivity in neural circuits. It offers a data-driven, systematic approach to decipher the genetic regulation of neural circuit formation. Nevertheless, the study is not without its limitations, which, in turn, highlight potential avenues for future research.
Firstly, the inherent linearity of the model facilitates the identification of key genes influencing neuronal connectivity but reduces the intricate, multi-stage process of synapse formation to a linear transformation. This simplification, while useful for gaining insights, does not fully capture the complexity of synapse formation. However, the bilinear model provides a solid theoretical foundation for developing more sophisticated, non-linear models. For instance, we could extend the bilinear model by incorporating kernels, possibly in the form of neural networks, to capture non-linear interactions between gene expressions and connectivity patterns (Figure 7). A manifestation of this non-linear approach is the “two-tower model”, in which each “tower” represents a deep neural network that learns a non-linear transformation of the input features [53, 54]. This model, widely used in contemporary recommendation systems, has demonstrated potency in capturing complicated user-item interactions.
Secondly, our model operates on the assumption that the transcriptomic profile of a neuron type largely determines its connectivity profile. However, other factors, such as neuronal activity, also influence synaptic connectivity, and our model does not account for these elements. Future models could incorporate activity-dependent variables and additional omics data, to capture a broader range of factors shaping neural circuits.
References
- [1]Connectome: How the Brain’s Wiring Makes Us Who We Are. Houghton Mifflin Harcourt
- [2]Establishment of axon-dendrite polarity in developing neuronsAnnual review of neuroscience 36:467–488
- [3]Synaptic Specificity, Recognition Molecules, and Assembly of Neural CircuitsCell 181:536–556
- [4]Neuronal cell-type classification: challenges, opportunities and the path forwardNature Reviews Neuroscience 18:530–546
- [5]Computational and analytical challenges in single-cell transcriptomicsNature Reviews Genetics 19:133–145
- [6]Serial block-face scanning electron microscopy to reconstruct three-dimensional tissue nanostructurePLOS biology 2
- [7]Connectomic reconstruction of the inner plexiform layer in the mouse retinaNature 500:168–174
- [8]High-contrast en bloc staining of neuronal tissue for field emission scanning electron microscopyNature Protocols 7:193–206
- [9]Matrix factorization techniques for recom-mender systemsComputer 42:30–37
- [10]Retinal bipolar cells: elementary building blocks of visionNat Rev Neurosci 15:507–519
- [11]The types of retinal ganglion cells: Current status and implications for neuronal classificationAnnual Review of Neuroscience 38:221–246
- [12]Eye smarter than scientists believed: Neural computations in circuits of the retinaNeuron 65:150–164
- [13]Cell types, circuits, computationCurrent Opinion in Neurobiology 21:664–671
- [14]Specification of synaptic connectivity by cell surface interactionsNature Reviews Neuroscience 17
- [15]The field of neurogenetics: where it stands and where it is goingGenetics 218
- [16]Bridging the Gap between Connectome and TranscriptomeTrends in Cognitive Sciences 23:34–50
- [17]Introduction to recommender systems handbookRecommender systems handbook 1:1–35
- [18]A survey of collaborative filtering techniquesAdvances in artificial intelligence, 2009
- [19]Bpr: Bayesian personalized ranking from implicit feedbackProceedings of the twenty-fifth conference on uncertainty in artificial intelligence :452–461
- [20]Type ii cadherins guide assembly of a direction-selective retinal circuitCell 158:793–807
- [21]Cadherin combinations recruit dendrites of distinct retinal neurons to a shared interneuronal scaffoldNeuron 99:1145–1154
- [22]Comprehensive classification of retinal bipolar neurons by single-cell transcriptomicsCell 166
- [23]Single-cell profiles of retinal neurons differing in resilience to injury reveal neuroprotective genesbioRxiv
- [24]the EyeWirers. Digital museum of retinal ganglion cells with dense anatomy and physiologyCell 173:1293–1306
- [25]the EyeWirers. Analogous convergence of sustained and transient inputs in parallel on and off pathways for retinal motion computationCell Reports 14:1–9
- [26]Unified classification of mouse retinal ganglion cells using function, morphology, and gene expressionCell Reports 40
- [27]Integrating single-cell transcriptomic data across different conditions, technologies, and speciesNature Biotechnology 36:411–420
- [28]Comprehensive Integration of Single-Cell DataCell 177:1888–1902
- [29]Factorized linear discriminant analysis and its application in computational biologyarXiv preprint
- [30]Detection of high variability in gene expression from single-cell RNA-seq profilingBMC Genomics 17
- [31]Comprehensive Identification and Spatial Mapping of Habenular Neuronal Types Using Single-Cell RNA-SeqCurr. Biol 28:1052–1065
- [32]Modular transcriptional programs separately define axon and dendrite connectivityeLife 8
- [33]Reconstruction of neocortex: Organelles, compartments, cells, circuits, and activityCell 185:1082–1100
- [34]Space–time wiring specificity supports direction selectivity in the retinaNature 509:331–336
- [35]The neuronal organization of the retinaNeuron 76:266–280
- [36]The functional diversity of retinal ganglion cells in the mouseNature 529:345–350
- [37]Transmembrane semaphorin signaling controls laminar stratification in the mammalian retinaNature 470:259–263
- [38]On and off retinal circuit assembly by divergent molecular mechanismsScience 342
- [39]Satb1 regulates contactin 5 to pattern dendrites of a mammalian retinal ganglion cellNeuron 95:869–883
- [40]Sidekick 2 directs formation of a retinal circuit that detects differential motionNature 524:466–470
- [41]Tbr1 instructs laminar patterning of retinal ganglion cell dendritesNature Neuroscience 21:659–670
- [42]g:Profiler — a web-based toolset for functional profiling of gene lists from large-scale experimentsNucleic Acids Research
- [43]g:Profiler: a web server for functional enrichment analysis and conversions of gene lists (2019 update)Nucleic Acids Research
- [44]Delta-protocadherins in health and diseaseProgress in Molecular Biology and Translational Science 116:169–192
- [45]Delta-protocadherins: organizers of neural circuit assemblySeminars in Cell & Developmental Biology 69:83–90
- [46]Regulation of neural circuit formation by protocadherinsCellular and Molecular Life Sciences 74:4133–4157
- [47]Adult mouse cortical cell taxonomy revealed by single cell transcriptomicsNature Neuroscience 19:335–346
- [48]Shared and distinct transcriptomic cell types across neocortical areasNature 563:72–78
- [49]A taxonomy of transcriptomic cell types across the isocortex and hippocampal formationCell 184:3222–3241
- [50]Network anatomy and in vivo physiology of visual cortical neuronsNature 471:177–182
- [51]Anatomy and function of an excitatory network in the visual cortexNature 532:370–374
- [52]A whole-brain monosynaptic input connectome to neuron classes in mouse visual cortexNature Neuroscience 26:350–364
- [53]Personalized embedding-based e-commerce recommendations at ebayarXiv preprint
- [54]Beijing. A dual augmented two-tower model for online large-scale recommendationKDD
Article and author information
Author information
Version history
- Preprint posted:
- Sent for peer review:
- Reviewed Preprint version 1:
- Reviewed Preprint version 2:
- Version of Record published:
Copyright
© 2023, Mu Qiao
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
- views
- 1,254
- downloads
- 102
- citation
- 1
Views, downloads and citations are aggregated across all versions of this paper published by eLife.