Sensory signals undergo substantial recoding when neural activity is relayed from sensors through pre-thalamic and thalamic nuclei to cortex. To explore how temporal dynamics and directional tuning are sculpted in hierarchical vestibular circuits, we compared responses of macaque otolith afferents with neurons in the vestibular and cerebellar nuclei, as well as five cortical areas, to identical three-dimensional translational motion. We demonstrate a remarkable spatio-temporal transformation: otolith afferents carry spatially aligned cosine-tuned translational acceleration and jerk signals. In contrast, brainstem and cerebellar neurons exhibit non-linear, mixed selectivity for translational velocity, acceleration, jerk and position. Furthermore, these components often show dissimilar spatial tuning. Moderate further transformation of translation signals occurs in the cortex, such that similar spatio-temporal properties are found in multiple cortical areas. These results suggest that the first synapse represents a key processing element in vestibular pathways, robustly shaping how self-motion is represented in central vestibular circuits and cortical areas.https://doi.org/10.7554/eLife.20787.001
Many sensory systems have been studied along their hierarchy, from primary receptor cells to cortical neurons. Such systematic analyses provide a foundation for comprehending the neural computations that convert early sensory activity into higher level constructs that underlie perception, action, and other cognitive functions. For the vestibular system, however, such analysis has been largely limited to reflex generation in the brainstem and cerebellum. On the other hand, there are multiple cortical areas that respond to vestibular stimuli, typically together with other sensory and motor signals (Grüsser et al., 1990; Bremmer et al., 2002; Fukushima et al., 2006; Klam and Graf, 2006; Chen et al., 2010, 2011a, 2011b, 2011c; Gu et al., 2006, 2016). How do the spatial and temporal properties of neurons in cortical areas differ from those in subcortical vestibular hubs?
Primary otolith afferents are spatially cosine-tuned and their temporal dynamics are broadly thought to encode translational acceleration (Fernández and Goldberg, 1976a, 1976b, 1976c; Angelaki and Dickman, 2000; Jamali et al., 2009). Neural response properties in the vestibular nuclei (VN), which receive the bulk of vestibular afferent projections outside the cerebellum (Barmack, 2003; Newlands and Perachio, 2003; Angelaki and Cullen, 2008), are different. It has been proposed that spatio-temporal convergence of otolith afferents onto central VN cells results in complex, non-cosine-tuned properties, where spatial and temporal coding might not always be multiplicatively separable (Angelaki, 1991, 1992b; Angelaki et al., 1992; Angelaki, 1993; Angelaki et al., 1993; Bush et al., 1993; Angelaki and Dickman, 2000; Dickman and Angelaki, 2002; Chen-Huang and Peterson, 2006, 2010). Some studies have supported this prediction in the vestibular brainstem (Angelaki et al., 1993; Bush et al., 1993; Angelaki and Dickman, 2000; Chen-Huang and Peterson, 2006, 2010), but direct comparisons with cortical responses (e.g., Chen et al., 2011c) have never been made.
Here, we use data obtained from transient translational displacements along multiple directions in three-dimensional (3D) space to compare the spatio-temporal response properties of neurons in the vestibular and rostral medial cerebellar nuclei (VN/CN; Liu et al., 2013) with responses in multiple cortical areas, including the parietoinsular vestibular cortex (PIVC; Chen et al., 2010), visual posterior sylvian area (VPS, a visual/vestibular convergent area just posterior to PIVC; Chen et al., 2011b), ventral intraparietal area (VIP; Chen et al., 2011c), dorsal medial superior temporal area (MSTd; Gu et al., 2006; Takahashi et al., 2007; Gu et al., 2010) and the frontal eye fields (FEF; Gu et al., 2016). In addition, we have also recorded from primary otolith afferent fibers from the vestibular subdivision of the eighth nerve in response to identical stimuli (Yu et al., 2015). We report a remarkable spatio-temporal transformation between otolith afferents and VN/CN cells, and this transformation determines the main response properties carried forward to cortical neurons.
The neural data set used to fit 3D spatio-temporal models consisted of the average temporal response profile (PSTH) of each neuron for 26 directions of translation corresponding to all possible combinations of azimuth and elevation angles in increments of 45° (Gu et al., 2006). The temporal waveform of the translational stimulus followed a Gaussian velocity profile (Figure 1, top), with corresponding biphasic acceleration (Figure 1, middle) and triphasic jerk (derivative of acceleration; Figure 1, bottom) components. We considered only cells with significant spatial and temporal response modulation, as detailed by Chen et al. (2011a) (see also Materials and methods). This inclusion criterion yielded a total of 27 otolith afferents (OA), 49 VN cells, 61 CN cells, 115 PIVC cells (from Chen et al., 2010), 66 VPS cells (Chen et al., 2011b), 139 MSTd cells (from Gu et al., 2006), 62 VIP cells (from Chen et al., 2011c) and 57 FEF cells (from Gu et al., 2016) (see Table 1).
Multiple models of varying complexity were fit to the PSTHs of each neuron (Figure 1). In its most general form, the standard model consisted of the sum of three response components, having temporal dynamics associated with velocity, acceleration and jerk. For each component, the temporal profile of the stimulus ( or ) was multiplied by a 3D spatial tuning function (, , cosine tuning with an offset; see Materials and methods) and a weight (, or ). This required four fitted parameters for each component. The sum of these three components was added to the resting discharge (), and a temporal delay term () was introduced (Figure 1). Thus, the maximum number of free parameters (‘VAJ’ model), was 14. For an easy comparison of the relative importance of the three temporal components, we also present the normalized weights (, and ), which are equal to the weights divided by ++.
Simpler models, consisting of either single or double component contributions were also fit to the responses of each cell, as described above, and included Acceleration-only (‘A’), Velocity-only (‘V’), Jerk-only (‘J’), Velocity+Acceleration (‘VA’), Velocity+Jerk (‘VJ’) and Acceleration+Jerk (‘AJ’) models. Because different models have different numbers of free parameters (6 parameters for the single-component models and 10 for the double-component models), the relative quality of the different model fits was assessed using the Bayesian Information Criterion (BIC; Schwarz, 1978).
Figure 2 illustrates example fits of the VAJ model for an OA (Figure 2A). The biphasic response PSTHs resembled the stimulus acceleration profile, with its amplitude being spatially modulated. Peak responses, with opposite signs, were observed during upward and downward motion. Furthermore, the cell responded weakly to stimuli with 0° elevation regardless of azimuth, corresponding to orthogonal directions relative to upward/downward motion. This response pattern is characteristic of cosine spatial tuning, where the response is modulated as a function of the cosine between the stimulus direction and the cell’s preferred direction.
For the example OA, the acceleration component had a large weight (Figure 2B, red) and partial R2 (Figure 2C, red); the jerk component had a lower, but still sizable, weight and partial R2 (blue), whereas the velocity component was negligible (Figure 2B–D). Accordingly, the AJ model was selected as best describing this cell’s response based on BIC analysis (even though Figure 2 illustrates the fits and parameters of the VAJ model). Note that the spatial tuning of acceleration and jerk were very similar (color maps in Figure 2E,F), indicating that spatial and temporal properties are separable (i.e. one does not depend on the other). This property is reflected in the cell’s separability index (Sep.I, Figure 2C, black; see Materials and methods) of 1. Figure 2G further summarizes the spatial tuning properties of all three components by plotting the their response offset against the amplitude of cosine tuning, each multiplied by the weight of the respective components. Since the acceleration and jerk responses are purely cosine-tuned, these components have zero offset while both the response offset and the cosine tuning of the velocity component are close to zero. The fitted temporal response components at three example stimulus directions are shown in Figure 2G,H,I (marked accordingly in the PSTHs of Figure 2A).
In contrast to otolith afferents, many central neurons exhibited distinct spatial tuning for the different temporal response components, as illustrated by data from an example VN neuron (Figure 3A–I). Here, all dynamic components contributed substantially to the cell’s response (Figure 3B,C), and the VAJ model had the best BIC. Unlike otolith afferents, this VN cell had a strong velocity modulation, which was positive along the preferred direction (close to downward; Figure 3D and H, green). However, the velocity component also exhibited a small positive (rather than negative) response during horizontal and upward motion directions. Thus, the velocity response component was positive in all directions (Figure 3D), which was modeled with a response offset greater than the amplitude of the cosine tuning (Figure 3G, green). The acceleration response was roughly cosine tuned, alternating from a positive response in downward movement directions to a negative response for upward movement directions (Figure 3E; see also Figure 3H–J, red); accordingly, its offset was close to zero (Figure 3G). Finally, the jerk response tuning was broad and mostly positive (Figure 3F; see also Figure 3G–I, blue).
Since the acceleration, velocity and jerk components had distinct spatial tuning, this VN neuron’s separability index (0.73) was much lower than that for OAs, and the interplay of these components created strikingly distinct (i.e. mono-, bi- and tri-phasic) PSTH shapes for different stimulus directions (Figure 3A). For example, a combination of positive velocity and acceleration responses (Figure 3G) created a largely monophasic profile during downward motion (Figure 3A, bottom). In contrast, the PSTH was biphasic during upward motion, which was essentially driven by acceleration and jerk (Figure 3I). Note that the jerk temporal modulation was most obvious at horizontal stimulus directions, when the cosine-tuned acceleration was minimal (Figure 3H). The fit of the VAJ model captures a large portion of this diversity in the shapes of PSTHs across stimulus directions.
The preferred directions (PDs) for velocity and acceleration were 28° apart (difference in 3D, with peaks at [azimuth, elevation] = [333°, 49°], and [3°, 74°], respectively; Figure 3D,E, ‘+’ in color contour plots). However, the jerk component had a nearly opposite preferred direction (at [azimuth, elevation] = [189°, −61°]; Figure 3F) and the difference with the preferred direction of the acceleration component was 167°. Note, however, that, given the strongly non-cosine-tuned properties of velocity and jerk (as captured by the non-zero offset parameters), PD was inadequate to fully characterize the spatial properties of most central cells.
Additional examples of cell responses and model fits (Figure 3—figure supplements 1–6 illustrate a diversity of non-linear response types encountered in central neurons. Remarkably, some neurons had either purely positive responses in all directions (Figure 3—figure supplements 1 and 3) or purely suppressive responses in all directions (Figure 3—figure supplement 5), whereas other cells responded only in a narrow range of motion directions (Figure 3—figure supplements 4 and 6). The VAJ model captured this variety of responses reasonably well, and therefore, an analysis of its parameters allowed us to draw quantitative conclusions from comparisons of cell tuning across neuronal populations.
Model fits were generally good (Figure 4A), with median R2 values of 0.78 for OA (CI=[0.67–0.88]), 0.65 for VN (CI=[0.59–0.73]), 0.64 for CN (CI=[0.57–0.68]), 0.71 for PIVC (CI=[0.67–0.75]), 0.67 for VPS (CI=[0.62–0.71]), 0.59 for MSTd (CI=[0.55–0.64]), 0.66 for VIP (CI=[0.59–0.66]) and 0.59 for FEF (CI=[0.52–0.66]). The percentage of neurons that were best fit by each of the 7 (V, A, J, VA, VJ, AJ and VAJ) models is shown in Figure 4B. Remarkably, there was a sharp contrast between OAs and central cells. Forty-eight percent of OAs were best fit by the A model and 33% by the AJ model, whereas only 19% were best fit by the VAJ model. This indicates that acceleration responses are ubiquitous in OAs, whereas half of them respond to jerk and only a few to velocity. In contrast, 6% of central cells were best fit by the V model, 20% and 10% by the VA and VJ model, and 60% by the VAJ model, indicating that velocity responses were very common in central neurons. For comparison with previous work (Chen et al., 2011c), we also tested a more complex model in which the non-linear function assumed the form of an exponential. We found that this model performed only marginally better (Figure 4—figure supplement 1); thus, we did not consider it further in this study.
The pattern of results seen in Figure 4B was also reflected in the partial R2 values (squared partial correlation coefficients, Figure 4C–E), which indicate how much each model component contributed to neural responses. In OAs, acceleration had high partial R2 values (median: 0.51, CI=[0.36–0.61]), while the jerk and velocity contributions were small (median partial R2: 0.07, CI=[0.02–0.13] and 0.03, CI=[0.02–0.06], respectively). In central neurons, the velocity contribution was greatest (median partial R2: 0.38, CI=[0.35–0.4]), followed by acceleration and jerk (median partial R2 : 0.23, CI=[0.21–0.25] and 0.14, CI=[0.13–0.15], respectively).
Next, we will summarize in detail the model fits of otolith afferents, followed by a quantitative description of model parameters in other brain areas.
As shown in Figure 4B, about half (n = 13) of OAs were best fit by the ‘A’ model (Figure 5, filled symbols), whereas the remainder (n = 14) were best fit by models of higher complexity (‘AJ’ and ‘VAJ’, open symbols in Figure 5). We found that these two groups had distinct firing properties. In particular ‘A’ afferents had a lower normalized coefficient of variation CV* (Goldberg et al., 1984) than ‘AJ/VAJ’ afferents (median 0.026, CI=[0.024–0.033] versus 0.06, CI=[0.04–0.08], Wilcoxon rank sum test, p=10−3; Figure 5A). We also found that the higher the CV* the higher the weights for velocity, acceleration and jerk (-red, -blue; Figure 5B, Spearman’s rank correlation p<10−2 for all three variables), as response gain increases with CV*. Importantly, the relative contributions of acceleration and jerk, as indicated by the normalized weights wA and wJ, (Figure 5C) were oppositely correlated with CV* (Spearman’s rank correlation, p<10−3 for both). Specifically, the normalized weight on acceleration declined with CV* while the normalized weight on jerk increased with CV*. In contrast, the contribution of velocity remained minimal (median = 0.06, CI=[0.05–0.08]) and did not correlate with CV* (Spearman’s rank correlation, p=0.75; Figure 5C, green).
The preferred directions of the acceleration and jerk components were always aligned for OAs, as illustrated by the clustering of the A-J preferred direction differences (Figure 5D, black) close to 0° (median = 7°, CI=[4-9]). Thus, the two temporal components that dominated otolith afferent responses were spatially aligned, which is consistent with separable spatio-temporal tuning. Although velocity weights were overall small, the preferred direction of the velocity component relative to acceleration or jerk components (V-A or V-J) showed a significant dependence on CV* (both r = 0.7, p<0.01, Spearman's rank correlation; Figure 5D). For afferents with the most regular firing rates, the small velocity components tended to be orthogonal to acceleration and jerk components (V-A and V-J preferred direction differences ~90°). In contrast, the small velocity components of irregular otolith afferent responses tended to have direction preferences opposite to those of acceleration and jerk components (V-A and V-J relative angle differences ~180°; Figure 5D). The presence of this velocity component in OAs has not been identified before using sinusoidal stimuli (see Discussion).
Identical models were fit to the responses of central neurons, with results summarized in Figures 6–9. We found large differences between the spatio-temporal properties of all central neurons, as compared to otolith afferents. The fitted velocity, acceleration and jerk component weights across areas are illustrated as ternary plots, which graphically depict the ratios of three variables that sum to a constant (in this case, the normalized weights, ++) as positions in an equilateral triangle (Figure 6). The value of each weight is 1 in one corner of the triangle and each weight decreases with increasing distance from this corner.
As described above, OAs (Figure 6A) respond predominantly to acceleration and display varying degree of sensitivity to jerk, while their response to velocity is minimal. Accordingly, they cluster between the A and J corners. The ternary plots of VN and CN neurons (yellow and orange, respectively) show a strikingly different pattern. Qualitatively, it is readily apparent that VN/CN cells carry substantially less acceleration signals, and much more velocity signals than otolith afferents. Quantitatively (see Table 2), the normalized acceleration weight (Figure 6E), which is high for otolith afferents, becomes significantly smaller for VN and CN neurons (p<10−8, Wilcoxon rank sum test). In parallel, the velocity weights (Figure 6D) are significantly greater for VN and CN neurons, as compared to otolith afferents (p<10−14, Wilcoxon rank sum test). There is little difference between jerk weights (Figure 6F) of VN/CN neurons and otolith afferents (p=0.03, Wilcoxon’s rank sum test). Remarkably, despite dramatic differences in the relative weighting of acceleration and velocity signals between central cells and otolith afferents, VN and CN have similar properties overall, and there is no significant difference between weight values for the two areas (Wilcoxon’s rank sum test, : p=0.75; ; p=0.47; : p=0.28).
As illustrated in Figure 6B,C and Figure 6D–F, the similarity in the relative weights among VN and CN neurons also extends to cortical areas (Table 2). The most noticeable difference among cortical areas is the slightly greater velocity weight in visual/vestibular multisensory areas (MSTd, VPS, VIP and FEF; median: 0.39, CI=[0.38–0.41]), as compared to those without optic flow responsiveness (VN, CN, PIVC; median 0.32, CI=[0.31–0.35]) (Figure 6D, cold vs. warm colors, respectively; p<10−9, Wilcoxon rank sum test). In fact, the normalized acceleration weight decreased, while the velocity weight increased from PIVC (median = 0.41 [0.38–0.44] CI; median = 0.33 [0.29 0.35] CI) to VIP (median = 0.39 [0.35–0.42] CI; median = 0.38 [0.34 0.4] CI) to MSTd (median = 0.28 [0.26–0.32] CI; median = 0.43 [0.4 0.45] CI), and these differences were significant (PIVC vs. MSTd: : p=3.10−9: p = 3.10−8; VIP vs. MSTd: : p = 4.10−4: p = 3.10−3; PIVC vs. VIP: : p=0.04; : p = 0.3, rank sum test). Yet, these differences are substantially smaller than the transformation of vestibular translation signals between otolith afferents and VN/CN neurons.
To further characterize the transformation of spatio-temporal response properties from afferents to cortex, Figure 7A shows distributions of the separability index. OAs that were fit with the AJ or VAJ models (n = 14) generally have very high separability indices (median 0.98, CI=[0.97–0.99]), indicating that their strongest temporal components have similar preferred directions and spatial tuning. The remaining OAs carry only an acceleration component, a property that automatically confers a high separability index (median = 0.99, CI = [0.98–0.99]). In contrast, central brain areas have overall lower separability indices (median 0.89, CI=[0.88–0.9]; Wilcoxon rank sum test, p<10−12). Some central cells are also spatio-temporally separable, but many others are not. For the latter neurons, each temporal component typically has a distinct directional tuning curve, resulting from a dependence of the temporal response profile on stimulus direction (e.g. Figure 3). Within central cells, we found a significant difference between the separability indices of VN/CN and cortical regions (median 0.82, [0.81 0.84] CI versus 0.9, [0.89 0.91] CI, Wilcoxon rank sum test, p<10−8). Differences in the separability index were much smaller among cortical areas (ANOVA, F1,4 = 2.51, p=0.04).
The preferred directions of neurons were widely scattered in all brain regions, such that the distributions did not differ significantly from uniform in any individual region (Figure 7—figure supplement 1). Figure 7B–D shows the color-coded distribution of the angular difference in 3D preferred direction for each pair of significant temporal components. Bold/italic labels below each bar graph indicate distributions that are/are not significantly different from a uniform (‘H1’) distribution. For comparison, a distribution that reflects purely aligned tuning (‘H2’) is also shown.
Here too, the largest differences are seen between otolith afferents and central cells. The preferred directions for acceleration and jerk are aligned in all afferents that are tuned to these components (Figure 7D; see also Figure 5D, black). Because few afferents are tuned to velocity (based on BIC analysis), the angles between the velocity component and the jerk or acceleration components were not significantly different from uniform overall (even though they show a strong dependence on CV*; Figure 5D).
In contrast to otolith afferents, central neurons respond predominantly to the velocity and acceleration components of the stimulus (Figure 4C,D; Figure 6). We found that these components have aligned spatial tuning for a large subpopulation of central neurons (Figure 7B, red); thus, the distributions were all significantly different from uniform (‘H1’). In addition, they were all also clearly different from a purely aligned distribution (Figure 7B–D, right bars titled ‘H2’). This finding is consistent with the wide spread of separability indices (Figure 7A). Overall, similar observations also hold for the distributions of V-J and A-J preferred direction differences (Figure 7C,D). Note that the small separability index of VN/CN neurons (Figure 4A) is in line with a greater dispersion of differences in preferred directions between A, V and J components in VN and CN neurons (color-coded in red in Figure 7B–D). For example, only 19% of preferred directions (VA, VJ and AJ angles pooled) were aligned within 30° of each other in VN/CN versus 33% in other central brain regions.
Large differences between otolith afferent and central responses were also observed when comparing other model parameters (Figure 8; see also Table 2). Otolith afferents had the smallest modulation amplitude (peak-to-trough amplitude of the spatio-temporal response curves, computed from the VAJ model fit) as compared to all other brain areas (median = 10 spikes/s, CI=[8-18] for OA versus 43 spikes/s, CI=[41-46] for all central neurons, Wilcoxon rank sum test: p=10−11; Figure 8A). Furthermore, modulation amplitude varied significantly across central brain regions (one-way ANOVA, F6,545 = 10, p=1.5*10−10). VN, CN and PIVC neurons were the most responsive (Figure 8A, warm colors), with aggregate modulation amplitude (median = 58 spikes/s, CI=[54-63]) greater than visual/vestibular multisensory areas VPS, MSTd, VIP and FEF (Figure 8A, cold colors; median: 35 spikes/s, CI=[33-39], Wilcoxon’s rank sum test: p<10−13). The reverse pattern of results was observed for the baseline response (one-way ANOVA across all brain regions: F7,571 = 74, p<10−75). FRo, which was greatest in otolith afferents (median = 72 spikes/s, CI=[54-90], Figure 8B, brown), intermediate in VN/CN (median = 45 spikes/s, CI=[38-49], Figure 8B, yellow/orange), and smallest in cortical areas (median = 12 spikes/s, CI=[11-13], Figure 8B). Thus, small gain modulation and high baseline firing rates in primary otolith afferents are converted into high gain modulation and low baseline firing rates in central brainstem, cerebellar and cortical neurons. Note, however, that this baseline firing rate could include a steady-state response to gravity (since animals were always oriented upright), which is unequivocal in otolith afferents but reduced across the population in brainstem, cerebellar and cortical neurons (Angelaki et al., 2004; Liu and Angelaki, 2009; Liu et al., 2011; Shaikh et al., 2005). An interesting picture also emerges when evaluating the spatial tuning, as summarized next.
As illustrated in Figures 2 and 3 and Figure 3—figure supplements 1–6, neuronal responses often showed non-linear spatial tuning that could be modeled by adding an offset to a cosine tuning function. The relative importance of offset and cosine-tuned responses are summarized in Figure 9. As illustrated in Figure 2, OAs mainly encode acceleration and jerk with cosine tuning. Accordingly, the offset parameters of the acceleration and jerk components cluster around zero for OAs (Figure 9B,C). In contrast, the weak velocity components of OAs display higher offsets (median offset parameter = 0.65, CI = [0.35–0.8]; Figure 9A).
Acceleration responses in central brain regions displayed more variable offsets, which nonetheless clustered around zero (median offset parameter = 0.08, CI = [0.06–0.11], Figure 9B). In absolute value, the cosine tuning amplitude was greater than the offset magnitude (i.e. ) for 88% of central neurons. These results indicate that acceleration responses tends to remain cosine-tuned throughout the brain.
In contrast, velocity responses usually displayed large positive offsets in all central brain regions (Figure 9A): the median offset parameter increased to 0.5 (CI = [0.47–0.52], rank sum test compared to : p=10−43; Figure 9—figure supplement 1A) and 50% of cells exhibited positive offsets greater than the amplitude of their cosine tuning component (i.e. ). Interestingly, a sizeable fraction (6%) of central cells exhibited negative offsets greater than their tuning (i.e. ), leading to omnidirectional inhibitory responses (as in Figure 3—figure supplement 5). Finally, we found that jerk responses in central neurons also exhibited positive offsets (median 0.49, CI=[0.46 0.52]; Figure 9C) comparable to velocity responses.
The VAJ model has been based on the main response components encountered in all brain areas. We also fitted neuronal responses with a PVAJ model that included a position component, in addition to velocity, acceleration and jerk. For each cell, we extracted the amplitude of the position response component (Figure 10A, inset, black) and the total peak to peak modulation (Figure 10A, inset, red) for each spatial direction. The maximum total modulation (across all spatial directions) is plotted versus the maximum position modulation in Figure 10A. We define the position ratio as the ratio of these two modulations. The diagonal and the two broken lines represent position ratios of 1:1, 1:3 and 1:10, respectively. Cells with a strong position response have a position ratio close to 1:1 and appear close to the diagonal, whereas cells with a weak position response appear far above the diagonal.
Many central neurons had significant position contributions (Figure 10A,B, filled symbols and bars): OA: 1/27 (4%), VN: 41/49 (84%), CN: 47/61 (77%), PIVC: 63/115 (55%), VPS: 51/69 (74%), MSTd: 108/139 (78%), VIP: 40/62 (65%), FEF: 45/57 (79%). Furthermore, the percentage of neurons with normalized position weights greater than 1:3 varied across areas (Figure 10B, gray-highlighted areas). In agreement with the findings of Chen et al. (2011c), the position contribution was larger in MSTd, as compared to PIVC and VIP (areas for which the distribution is significantly different from MSTd are indicated by stars; *p<0.01; **p<0.001; two-sample Kolmogorov-Smirnoff tests). In addition to MSTd (cyan), we also found prominent position responses in VN (yellow), CN (orange) and FEF (black). The spatial properties of the position component were generally similar to those of the velocity component (shown for MSTd in Figure 10C,D), showing both inseparability and non-linear spatial tuning. Note, however, that future studies should record pre- and post-stimulus firing rates for a longer duration, such that these position responses can be quantified more reliably.
To understand the spatio-temporal processing of translational motion along the hierarchy of the vestibular system, from primary sensory neurons to subcortical and central cortical structures, we have performed a systematic comparison of spatio-temporal response properties to identical 3D naturalistic stimuli. Our analysis has revealed remarkably different spatio-temporal properties between central neurons and primary otolith afferents. Whereas otolith afferents encode mostly linear acceleration, central neuron responses are dominated by components related to stimulus velocity. Furthermore, central neurons also encode jerk (the derivative of linear acceleration), as well as position (displacement). However, this position response component needs to be further investigated in future studies using stimuli that allow a precise quantification of the pre-stimulus and post-stimulus firing rates (see Materials and methods).
The observed increase in velocity contribution to the spatio-temporal properties of central cells, as compared to otolith afferents, was accompanied by the presence of non-linearities. Whereas the linear acceleration component of central cells continued to be, like otolith afferents, mostly cosine-tuned, the velocity component included substantial omni-directional contributions, and jerk had intermediate properties. Thus, central neurons, like afferents, continued to carry information about the direction of linear acceleration, but they also encoded another type of signal: the presence or absence of translational motion without regard for direction. This component, which was strong in both the cortex and vestibular/cerebellar nuclei, cannot be characterized with the steady-state sinusoidal stimuli that have traditionally been the stimuli of choice in the vestibular system.
In contrast to the large differences in the types of signals encoded by otolith afferents versus their target cells in the vestibular nuclei, we found only moderate differences in the relative contributions of velocity, acceleration and jerk among central subcortical (vestibular and cerebellar nuclei) and cortical neurons. However, there were systematic differences in both the baseline firing and amplitude of stimulus-driven responses. Otolith afferents and vestibular/cerebellar nuclei neurons had high baseline firing compared to cortical neurons. Furthermore, response amplitude to an identical stimulus was lowest in otolith afferents and greatest in vestibular/cerebellar nuclei, PIVC and VPS – the areas known to form the core of the central vestibular system (Angelaki et al., 2009; Lopez and Blanke, 2011). In contrast, response amplitude is lower in MSTd, VIP and FEF, which are visual/vestibular multisensory areas. This result is consistent with Chen et al. (2011c), who speculated that the vestibular pathway might go from OA to VN/CN to PIVC to VIP to MSTd. However, although modulation strength remains high in VN/CN and PIVC, it is nevertheless reduced in multisensory visual areas MSTd, VIP and FEF. Note that, although otolith afferents have low peak-to-trough modulation, their information content is high because of the high firing rate regularity (Jamali et al., 2013).
It is important to point out that Chen et al. (2011c) compared neural responses in a more limited number of cortical areas during motion within a single plane and reported a gradual shift from acceleration dominance in PIVC to velocity dominance in MSTd, as well as stronger position contributions in MSTd than PIVC or VIP. The present findings are consistent with both conclusions. However, when the full 3D properties are considered in the present analyses, we find that the spatio-temporal tuning properties of individual cortical areas also exhibit remarkable similarities. Note also that the model of Chen et al. (2011c) did not include a jerk component, which may also contribute to the results. Purkinje cells in the cerebellar vermal lobules X and IX, which were also previously fitted using the planar model (Yakusheva et al., 2013), also showed similar properties to central brain areas when analyzed using the full 3D model used here (unpublished). In summary, these results suggest that the signal transformation that takes place between sensory afferents and their targets in the brainstem and cerebellum represents a key processing element in the vestibular system. Using identical passive stimuli, cortical areas show moderate differences, as compared to the response properties in the brainstem and cerebellum.
Understanding the central processing of translation signals is complicated for at least two reasons. First, afferents from a single otolith organ differ in both spatial and temporal response properties (Fernández and Goldberg, 1976a, 1976b, 1976c; Angelaki and Dickman, 2000; Yu et al., 2012; Jamali et al., 2013) and extensive convergence in the vestibular nuclei (Kushiro et al., 2000; Straka et al., 2002; Uchino et al., 2001; Uchino and Kushiro, 2011) creates complex spatio-temporal convergence properties (Angelaki et al., 1992; Bush et al., 1993; Angelaki and Dickman, 2000; Dickman and Angelaki, 2002; Chen-Huang and Peterson, 2006, 2010).
Second, otolith afferents cannot distinguish tilt relative to gravity from translational acceleration. However, this tilt/translation ambiguity has been resolved in the cortex, where neurons that are selective for inertial (translational) accelerations are found (PIVC: Liu et al., 2011; MSTd: Liu and Angelaki, 2009). This property of cortical neurons is likely inherited from subcortical signals, as there is growing evidence that this computation is implemented through otolith/canal convergence in the brainstem and cerebellum (VN/CN: Angelaki et al., 2004; Liu et al., 2013; caudal cerebellar vermis: Yakusheva et al., 2007; Laurens et al., 2013a, 2013b). It is thus possible that the large differences in spatio-temporal response properties that we have found between otolith afferents and VN/CN neurons arise from neural computations that resolve the tilt/translation ambiguity, a vital and critically important function for spatial orientation.
Otolith afferents have been previously characterized in terms of preferred direction in three-dimensions (Fernández and Goldberg, 1976a; Tomko et al., 1981; Yu et al., 2012) and response dynamics (Anderson et al., 1978; Fernández and Goldberg, 1976b; Goldberg et al., 1990; Si et al., 1997; Angelaki and Dickman, 2000; Purcell et al., 2003; Jamali et al., 2009). These two properties are separable, with the activity of each afferent fiber being determined by the product of a temporal ‘transfer’ function and a cosine-tuned spatial function (Fernández and Goldberg, 1976a, 1976b, 1976c). As a result of this separability, otolith afferents have the same response dynamics along different spatial directions.
It has long been recognized that otolith afferents differ in their response dynamics based upon the regularity of their spontaneous discharge (Anderson et al., 1978; Angelaki and Dickman, 2000; Blanks and Precht, 1978; Fernández and Goldberg, 1976b, 1976c; Goldberg et al., 1990; Jamali et al., 2009, 2013; Purcell et al., 2003). Indeed, we found that normalized jerk weights for otolith afferents increased with CV*, whereas acceleration weights decreased with CV*. Thus, the most regular afferents encoded pure acceleration, whereas more irregular afferents encoded mixtures of acceleration and jerk. These findings are consistent with previous studies reporting phase leads and gain advances in less regularly-firing afferents during sinusoidal stimulation (Anderson et al., 1978; Angelaki and Dickman, 2000; Fernández and Goldberg, 1976b; Goldberg et al., 1990; Jamali et al., 2013; Purcell et al., 2003; Si et al., 1997). Less regular-firing afferents also included a small but consistent velocity component, whose preferred direction relative to that of the jerk and acceleration components is correlated with CV*. Such a response component has not been described previously using sinusoidal stimulation.
Because VN neurons receive extensive convergence from otolith afferents that vary in their 3D spatial and temporal properties, a computational challenge arises. How would the properties of central neurons transform this information into signals that can be used by the rest of the brain? Theoretical analysis of such convergence showed that central neurons receiving otolith afferent convergence should, in general, exhibit spatio-temporal convergence properties, where spatial and temporal coding might not necessarily be separable; this results in complex spatial (non-cosine-tuned) properties, where central neurons carry different proportions of velocity, acceleration and jerk signals along different spatial directions (Angelaki, 1992a, 1992b, 1992c; Schor and Angelaki, 1992; Angelaki et al., 1993; Angelaki and Dickman, 2000). Evidence for simultaneous coding of both translational acceleration and jerk was indeed demonstrated in VN responses of decerebrate rodents (Angelaki et al., 1993).
It took several more years until some experimental evidence for spatio-temporal convergence properties was provided for neurons in the primate VN (Angelaki and Dickman, 2000; Chen-Huang and Peterson, 2006, 2010) and CN (Shaikh et al., 2005). However, to our knowledge, all previously published studies of responses of VN/CN neurons to translation used sinusoidal stimuli. There are many reasons why such stimuli are problematic. First, VN and CN responses are characterized by non-minimum phase properties, i.e., phase and gain changes do not parallel each other, thus complicating identification of temporal dynamics using sinusoids (Angelaki and Dickman, 2000; Dickman and Angelaki, 2002; Shaikh et al., 2005). Second, responses proportional to the temporal derivative and integral of translational acceleration cannot be easily separated using sinusoidal stimuli (e.g., a phase difference of 90° relative to acceleration could reflect either velocity coding in one direction or jerk coding in the opposite direction). The traditional approach of testing sinusoidal responses at different frequencies to distinguish velocity from jerk contributions using linear systems analyses is complicated by the presence of non-minimum phase properties, thus making system identification challenging (Angelaki and Dickman, 2000; Dickman and Angelaki, 2002; Shaikh et al., 2005). Furthermore, if indeed response dynamics vary with stimulus direction, a complete characterization of the spatio-temporal properties requires testing cells at multiple frequencies and multiple directions in 3D (Chen-Huang and Peterson, 2006, 2010). Third, non-linearities in spatio-temporal tuning, as reported here (see also Chen et al., 2011c; Massot et al., 2012) can violate the assumptions of traditional linear systems analysis, thus rendering previous conclusions based on steady-state frequency analysis problematic.
These limitations have been circumvented here, for the first time, by the use of 3D transient stimuli and a comparison of neural responses in multiple subcortical and cortical areas. Importantly, in contrast to Chen et al. (2011c), who fit cortical responses during translation in a single plane, we have developed models that characterize cell responses in 3D. With the exception of Chen-Huang and Peterson (2010), who characterized 3D responses in the VN using sinusoidal stimuli (see limitations above), nothing is known about spatio-temporal convergence in any brain region using transient stimuli. Yet, as previously also emphasized by Chen-Huang and Peterson (2010), understanding the full extent of the underlying spatio-temporal computations is only achievable when data are analyzed in 3D, and conclusions drawn by characterizing responses along a single direction or within a single plane can be rather incomplete. Our current findings have provided the first such quantification of 3D spatio-temporal convergence properties using transient stimuli along many spatial directions.
In contrast to cosine-tuning of otolith afferents, central neuron responses often show strongly nonlinear and rectified responses to both translational velocity and jerk. Most central cells show omni-directional tuning components. The existence of broad or uniform tuning, as opposed to cosine tuning, suggests that many afferents converge onto one central neuron. Indeed, broadly tuned neurons respond to motion in directions that are orthogonal to their PD. In contrast, in OA, all directions that are orthogonal to the PD are null directions.
These results suggest that the transformation of responses from sensory afferents to brainstem neurons represents a key processing element, possibly because it is coupled to the resolution of a tilt/translation ambiguity that represents a critically important function for spatial orientation. Although there is no direct evidence yet for such a link, theoretical solutions to the ambiguity problem require strong non-linearities (Borah et al., 1988; Merfeld, 1995; Glasauer and Merfeld, 1997; Merfeld et al., 1999; Angelaki et al., 1999, 2004; Laurens and Droulez, 2007; Laurens and Angelaki, 2011).
The functional significance of the observed spatio-temporal transformation that exists between otolith afferents and VN/CN cells, as well as all other cortical regions examined, remains to be determined. Still, it is striking that all central neurons simultaneously carry velocity, acceleration and jerk signals, with the relative contributions of each component depending on stimulus direction. Remarkably, we observed moderate further spatio-temporal processing (other than differences and response amplitude and baseline firing rate) beyond the VN/CN, such that roughly similar properties characterize all the cortical representations of vestibular translation signals examined to date.
Previous studies (Rigotti et al., 2013; Fusi et al., 2016) have proposed a role for the mixed non-linear responses widely observed in prefrontal cortex. It has been proposed that mixed selectivity plays an important computational role: high-dimensional representations with mixed selectivity allow a simple linear readout to generate a diverse array of potential responses (Fusi et al., 2016). In contrast, representations based on highly specialized neurons are low dimensional and may preclude a linear readout from generating several responses that depend on multiple task-relevant variables. Complex non-linear operations can be performed by simple linear summations of non-linear neurons, and in turn linear summation may be easily learned by neuronal networks. Thus, a rich variety of non-linearly transformed signals may facilitate the learning of complex computations.
It is possible that the described spatio-temporal transformations in VN/CN might reflect the need to perform non-linear, three-dimensional spatio-temporal operations necessary to implement an internal model of head motion (Merfeld, 1995; Laurens and Angelaki, 2011), whose influence has been documented in subcortical neurons (Angelaki et al., 2004; Yakusheva et al., 2007; Laurens et al., 2013a, 2013b). Complex spatio-temporal representations of movement may be passed on to cortical (Liu and Angelaki, 2009; Liu et al., 2011) neurons to subserve their respective roles in spatial cognition. This hypothesis remains to be tested in future experiments.
We include data from a total of 19 rhesus monkeys (Macaca mulatta). Each animal was chronically implanted with an eye coil, a head-restraint ring, and a plastic recording grid that contains an array of holes through which guide tubes were passed for extracellular electrophysiological recordings. All surgical and experimental procedures were approved by the Institutional Animal Care and Use Committee at Washington University and Baylor College of Medicine and were performed in accordance with institutional and NIH guidelines.
The data sample analyzed in the present study includes neurons recorded from multiple cortical areas (PIVC VIP, VPS, MSTd and FEF) using epoxy-coated tungsten microelectrodes (1–2 MΩ). Basic methodology and neural response properties of these cortical areas can be found in previous publications: PIVC: Chen et al. (2010), VPS: Chen et al. (2011a), VIP: Chen et al. (2011b), MSTd: Gu et al. (2006), 2010; Takahashi et al. (2007) and FEF: Gu et al. (2016). In order to compare with cortical neurons, we also recorded new data from ‘vestibular-only’ neurons without eye movement-related activity in the VN/CN (5–7 MΩ impedance electrodes) and otolith afferents (18–20 MΩ) using identical stimulation protocols (see below). To localize areas VN and CN, we first identified the abducens nuclei bilaterally based on their characteristic burst-tonic activity during horizontal eye movements (for details see Meng et al., 2005; Liu et al., 2013). The vestibular nerve was isolated beneath the auditory meatus as it entered the brain, as detailed in previous publications (Haque et al., 2004; Yu et al., 2012, 2015). Note that the VN/CN sample in the current study also includes some of the same neurons included in Liu et al. (2013). Samples of at least 40 neurons per region were collected; a sample size of 27 OA was considered sufficient due to the homogeneity of responses across OA.
Monkeys were seated comfortably in a primate chair, which was secured to a six-degree-of-freedom motion platform (MOOG 6DOF2000E). We examined each cell's 3D spatio-temporal tuning by recording neural responses while the animal was translated along each of 26 directions spaced evenly within a sphere. This stimulus set includes all combinations of movement vectors having eight different azimuth angles (0, 45, 90, 135, 180, 225, 270 and 315°; forward, backward, leftward and rightward movements correspond to 90, 270, 0 and 180° respectively) and three different elevation angles, 0° (the stereotaxic horizontal plane) and ±45°, for a subtotal of 8×3 = 24 directions, as well as two additional movement vectors with elevation angles of −90°and 90° corresponding to upward and downward directions, respectively. Each movement trajectory consisted of a Gaussian velocity profile (0.2 s standard deviation) with corresponding biphasic acceleration and triphasic jerk profiles. The total displacement was 13 cm and the peak acceleration was 0.1 g. The frequency content of this stimulus is illustrated in Figure 1B. All movements originated from the center of the movement range of the motion platform, and the platform returned to this starting position during the 2 s inter-trial interval. Stimuli were presented in random order within a single block of trials (at least five repetitions).
Data (neural activity and the translational acceleration stimulus) were collected either in complete darkness or during fixation of a central, head-fixed target in an otherwise dark room. Previous findings have established that otolith afferents, VN, CN and PIVC neurons are not sensitive to visual stimuli or eye movements (for details see Bryan and Angelaki, 2009; Liu et al., 2013; Chen et al., 2010). For MSTd, VPS, VIP and FEF neurons, which are sensitive to eye movements and/or visual motion, we have previously verified that vestibular translation responses during fixation are similar to those in complete darkness (details and comparison figures can be found in: Gu et al., 2006; Takahashi et al., 2007 and Chowdhury et al., 2009 for MSTd, Chen et al., 2011a for VPS, Chen et al., 2011b for VIP and Gu et al., 2016 for FEF).
Quantitative data analyses were performed off-line using custom-written scripts in Matlab (MathWorks), available at the following adress: https://github.com/JeanLaurens/Spatiotemporal_Dynamics. Peristimulus time histograms (PSTHs) were constructed for each direction of translation using 25 ms time bins and were smoothed with a Gaussian kernel (σ = 100 ms). We applied both temporal response modulation and space-time structure criteria to determine if a cell had responses strong enough to be included in the present analyses (see Chen et al. (2011a) for details). For the temporal response criterion, we first found the peak (trough) time of the PSTH for each stimulus direction. Using the peak (trough) time, we obtained a distribution of response values from the PSTHs for each trial at that point in time. We then compared this peak (trough) distribution with the distribution of values obtained from baseline firing using the Wilcoxon signed-rank test. If responses to two neighboring stimulus directions (45° apart) had significantly (p<0.01) different peak (trough) distributions compared to their own baseline distributions, then the cell passed the temporal response modulation criterion. For the space-time structure criterion, we performed a two-way ANOVA on the full set of single-trial PSTHs, with space and time as factors (Chen et al., 2011c). Cells were considered to have space-time structure when both factors as well as their interaction were highly significant (p<0.001).
To ensure robustness, non-parametric statistics were used and median results were reported whenever it was possible. Confidence intervals of medians were computed based on bootstrap resampling.
To characterize and quantify the spatiotemporal dynamics of vestibular responses, trial-averaged PSTHs for each direction of translation were fit with spatiotemporal functions of varying complexity. In its simplest single-component form, the fitted function consists of a spatial tuning component, and a temporal response profile , multiplied together to obtain the spatiotemporal response function (Figure 1):
where is the response amplitude and is the baseline firing rate. Note that and are normalized and unitless, such that and are both expressed in spikes/s. Other parameters are defined below.
The spatial tuning is modelled as a cosine function fed through a non-linear function . The cosine tuning function depends on parameters (preferred azimuth) and (preferred elevation):
where maps the unit vector in spherical coordinates to Cartesian coordinates (Mardia and Jupp, 1999). Because is equal to the cosine of the angle between and , it has a value of 1 at the preferred direction (PD), a value of −1 at the opposite direction (called anti-preferred direction) and a value of 0 in any direction orthogonal to the PD (called ‘an orthogonal direction’).
The non-linear function, generalizes the spatial tuning from a pure cosine function by adding an offset parameter, . It takes the form
The offset parameter, , ranging from −1 to 1, allows us to model omnidirectional response components, while the term represents cosine tuning as a function of motion direction. Note that scaling by ensures that the function is normalized, such that for all in and for at least one value of .
The temporal profile, , is defined as follows:
In Equations 4–6, (temporal delay) is a fitted parameter, whereas (temporal Gaussian width) is set by the stimulus ( = 0.2 s). The operator indicates that the temporal profiles were normalized so that the difference between the maximum and the minimum values of each profile is 1 (Figure 1, blue, red and green waveforms, respectively).
The simplest spatio-temporal model consisted of Velocity-only (‘V’ model), Acceleration-only (‘A’ model) or Jerk-only (‘J’ model) terms based on Equation 1, where the temporal profile was given by Equations 4, 5 or 6, respectively. The most general model was the Velocity+Acceleration+Jerk (VAJ) model, given by the following equation:
where , and are the velocity, acceleration and jerk components, respectively. Each of the three spatial tuning functions (has its own set of parameters (and , respectively), and , and are the respective weights on each component. Note that a single delay parameter is used in the VAJ model. Additional models of intermediate complexity—Velocity + Acceleration (‘VA’), Velocity + Jerk (‘VJ’), Acceleration + Jerk (‘AJ’)—were also tested. The 1-component models (‘V’, ‘A’ or ‘J’) have six free parameters, the two-component models (‘VA’, ‘VJ’ or ‘AJ’) have 10 free parameters, and the three-component model (‘VAJ’) has 14 parameters.
How well each model fit the neural responses was quantified as the proportion of variance accounted for by the model (R2), and was computed by regressing the responses of each neuron against the values of each fitted function (across the 26 heading directions and the entire 2 s response profile). To evaluate the best model while accounting for the number of model parameters, we used the Bayesian Information Criterion (BIC), defined as:
where is the residual sum of squares, is the number of data points (considering that the response profile was 2 s long and filtered by a Gaussian kernel with σ = 100 ms, we assumed that the data amounted to 10 independent time points per profile and therefore n = 260) and is the number of function parameters (Konishi and Kitagawa, 2008). The best model based on this criterion is the one with the lowest BIC value.
The importance of each model component was also assessed by computing its partial coefficient of correlation given the two other components. For instance, the partial correlation coefficient of the A component RA|VJ reflects how much variance is explained by adding the A component to the VJ model and is computed accorded to:
The VAJ model includes a separate spatial tuning curve for each dynamic component (acceleration, velocity and jerk). However, for certain cells, the dynamic components have an identical spatial tuning; a property called ‘spatio-temporal separability’. We tested if the tuning of cells was separable by fitting an additional ‘Separable VAJ’ model (which has eight parameters) that includes an identical spatial tuning curve for all components. This model was used to compute a ‘separability index, (Sep I)’:
In order to validate our approach, we simulated 1000 Poisson-spiking neurons, divided equally among each model type, and verified that fitting the standard model could satisfactorily retrieve the spatio-temporal tuning of neurons.
Finally, we also tested a four-component (PVAJ) model (18 parameters), in which a position (integral of velocity) component was added to the other three terms (velocity, acceleration and jerk). However, the recorded data used to evaluate these models were limited to at most 200 ms following motion offset, which makes the contribution of position modulation hard to evaluate reliably. For this reason, we focus mainly on the findings of the VAJ model (but see Figure 10).
Dynamic polarization vector of spatially tuned neuronsIEEE Transactions on Biomedical Engineering 38:1053–1060.https://doi.org/10.1109/10.99068
Spatio-temporal convergence (STC) in otolith neuronsBiological Cybernetics 67:83–96.https://doi.org/10.1007/BF00201805
Vestibular neurons encoding two-dimensional linear acceleration assist in the estimation of rotational velocity during off-vertical axis rotationAnnals of the New York Academy of Sciences 656:910–913.https://doi.org/10.1111/j.1749-6632.1992.tb25292.x
A model for the characterization of the spatial properties in vestibular neuronsBiological Cybernetics 66:231–240.https://doi.org/10.1007/BF00198476
Vestibular system: the many facets of a multimodal senseAnnual Review of Neuroscience 31:125–150.https://doi.org/10.1146/annurev.neuro.31.060407.125555
Central vestibular system: vestibular nuclei and posterior cerebellumBrain Research Bulletin 60:511–541.https://doi.org/10.1016/S0361-9230(03)00055-8
Functional characterization of primary vestibular afferents in the frogExperimental Brain Research 25:369–390.https://doi.org/10.1007/bf00241728
Optimal estimator model for human spatial orientationAnnals of the New York Academy of Sciences 545:51–73.https://doi.org/10.1111/j.1749-6632.1988.tb19555.x
Visual-vestibular interactive responses in the macaque ventral intraparietal area (VIP)European Journal of Neuroscience 16:1569–1586.https://doi.org/10.1046/j.1460-9568.2002.02206.x
Optokinetic and vestibular responsiveness in the macaque rostral vestibular and fastigial nucleiJournal of Neurophysiology 101:714–720.https://doi.org/10.1152/jn.90612.2008
Macaque parieto-insular vestibular cortex: responses to self-motion and optic flowJournal of Neuroscience 30:3022–3042.https://doi.org/10.1523/JNEUROSCI.4029-09.2010
Convergence of vestibular and visual self-motion signals in an area of the posterior sylvian fissureJournal of Neuroscience 31:11617–11627.https://doi.org/10.1523/JNEUROSCI.1266-11.2011
Representation of vestibular and visual cues to self-motion in ventral intraparietal cortexJournal of Neuroscience 31:12036–12052.https://doi.org/10.1523/JNEUROSCI.0395-11.2011
Does the middle temporal area carry vestibular signals related to self-motion?Journal of Neuroscience 29:12020–12030.https://doi.org/10.1523/JNEUROSCI.0004-09.2009
Vestibular convergence patterns in vestibular nuclei neurons of alert primatesJournal of Neurophysiology 88:3518–3533.https://doi.org/10.1152/jn.00518.2002
Why neurons mix: high dimensionality for higher cognitionCurrent Opinion in Neurobiology 37:66–74.https://doi.org/10.1016/j.conb.2016.01.010
Modelling three-dimensional vestibular responses during complex motion stimulationIn: M Fetter, T Haslwanter, H Misslisch, editors. Three-Dimensional Kinematics of Eye, Head and Limb Movements. Amsterdam: Harwood academic. pp. 387–398.
Spatial tuning and dynamics of vestibular semicircular canal afferents in rhesus monkeysExperimental Brain Research 155:81–90.https://doi.org/10.1007/s00221-003-1693-0
Response of vestibular nerve afferents innervating utricle and saccule during passive and active translationsJournal of Neurophysiology 101:141–149.https://doi.org/10.1152/jn.91066.2008
Saccular and utricular inputs to single vestibular neurons in catsExperimental Brain Research 131:406–415.https://doi.org/10.1007/s002219900312
The functional significance of velocity storage and its dependence on gravityExperimental Brain Research 210:407–422.https://doi.org/10.1007/s00221-011-2568-4
Computation of linear acceleration through an internal model in the macaque cerebellumNature Neuroscience 16:1701–1708.https://doi.org/10.1038/nn.3530
Choice-related activity and correlated noise in subcortical vestibular neuronsNature Neuroscience 16:89–97.https://doi.org/10.1038/nn.3267
The thalamocortical vestibular system in animals and humansBrain Research Reviews 67:119–146.https://doi.org/10.1016/j.brainresrev.2010.12.002
Pursuit--vestibular interactions in brain stem neurons during rotation and translationJournal of Neurophysiology 93:3418–3433.https://doi.org/10.1152/jn.01259.2004
Modeling the vestibulo-ocular reflex of the squirrel monkey during eccentric rotation and roll tiltExperimental Brain Research 106:123–134.https://doi.org/10.1007/BF00241362
Responses of gerbil utricular afferents to translational motionExperimental Brain Research 152:317–322.https://doi.org/10.1007/s00221-003-1530-5
The algebra of neural response vectorsAnnals of the New York Academy of Sciences 656:190–204.https://doi.org/10.1111/j.1749-6632.1992.tb25209.x
Properties of cerebellar fastigial neurons during translation, rotation, and eye movementsJournal of Neurophysiology 93:853–863.https://doi.org/10.1152/jn.00879.2004
Response properties of pigeon otolith afferents to linear accelerationExperimental Brain Research 117:242–250.https://doi.org/10.1007/s002210050219
Patterns of canal and otolith afferent input convergence in frog second-order vestibular neuronsJournal of Neurophysiology 88:2287–2301.https://doi.org/10.1152/jn.00370.2002
Responses to head tilt in cat eighth nerve afferentsExperimental Brain Research 41:216–421.https://doi.org/10.1007/BF00238878
Spatiotemporal properties of optic flow and vestibular tuning in the cerebellar nodulus and uvulaJournal of Neuroscience 33:15145–15160.https://doi.org/10.1523/JNEUROSCI.2118-13.2013
Jennifer L RaymondReviewing Editor; Stanford University, United States
In the interests of transparency, eLife includes the editorial decision letter and accompanying author responses. A lightly edited version of the letter sent to the authors after peer review is shown, indicating the most substantive concerns; minor comments are not usually included.
Thank you for submitting your article "Evolution of spatiotemporal dynamics from otolith sensory afferents to cortex" for consideration by eLife. Your article has been reviewed by three peer reviewers, one of whom, Jennifer L Raymond (Reviewer #1), is a member of our Board of Reviewing Editors and the evaluation has been overseen by David Van Essen as the Senior Editor.
The reviewers have discussed the reviews with one another and the Reviewing Editor has drafted this decision to help you prepare a revised submission.
This paper characterizes the spatial and temporal coding properties of neurons throughout the vestibular system hierarchy. This work may be unprecedented in its comprehensive and systematic characterization of neurons at so many different levels of a sensory system, using a common set of stimuli and analysis methods. 579 neurons were recorded in 8 subcortical and cortical regions carrying vestibular signals (otolith afferents, vestibular and cerebellar nuclei, and 5 cortical areas: PIVC, VIP, VPS, MSTd, FEF) during transient, translational vestibular stimuli delivered in 24 directions in 3D space. Neural responses were fit with a model that included velocity, acceleration and jerk, and linear or nonlinear spatio-temporal combinations of these quantities. The analysis is principled, systematic, uniformly applied across brain areas, and validated using simulated neurons with defined coding properties. The authors find that whereas there are clear differences in response selectivity between of primary otolith afferents and central neurons, responses of neurons in vestibular and cerebellar nuclei and five distinct cerebrocortical areas are remarkably similar to each other. The finding of similar spatio-temporal coding properties in all brain regions beyond the otolithic afferents suggests that a key step in the processing of vestibular information is performed at the first synapse. Overall, the work is sound and carefully executed. It is impressive in its completeness, providing a comprehensive picture of vestibular signal transformations across brain regions. The central findings provide a valuable resource for the vestibular field, and provide the opportunity for useful comparisons across sensory modalities. Although generally enthusiastic, the Reviewers had several suggestions for additional analysis, and for how the presentation of the Results and Discussion could be improved to increase the impact of this impressive body of research.
1) The main conclusion, that the big signal transformation happened at the first processing stage and that all the higher vestibular areas are more similar in their coding properties, is well supported. However, the difficulty for the reader to draw any conclusion beyond this creates an impression that not much was gained from the massive effort invested. A more nuanced treatment of differences between central brain areas seems merited, and would do the rich data set more justice. Some specific examples, raised by the Reviewers:
a) A previous study from this group (Yakusheva et al., J Neurosci 2013) had shown that vestibular responses of Purkinje cells in the nodulus/ ventral uvula were similar to those of MSTd, but not PIVC or VIP neurons. Purkinje cells are not recorded here, but it would be useful to cite this previous study to provide context for the current results, especially since the take-home message appears to be quite different (MSTd and Purkinje neurons were similar to each other but different from PIVC and VIP neurons in Yakusheva et al. 2013, but in the current manuscript the argument is that all central areas are more or less equivalent).
b) The ternary plots in Figure 6 seem to emphasize similarities between regions. For example, VIP and MSTd are plotted together in C, and this and the superposition of the envelopes of all of the data in D obscures the fact that these two regions actually have quite distinct velocity weights. Similarly, there are many more neurons in 6C (representing MSTd, VIP, FEF) that have 0 acceleration weights (fall along the bottom of the triangle) than in the other plots/ regions, but the envelope representation in 6D obscures that fact. Comparing this (Figure 6) approach to data visualization vs. the approach in Figure 10, it is not clear to me whether different conclusions about variability across regions are not reached partly because of differences in data analysis/ presentation. In Figure 10, where actual histograms are presented for position components for each region, the conclusion reached is quite different – that only MSTd has a sizable position component. Would we more easily pick out differences across regions in the velocity/ acceleration/ jerk weights of Figure 6 if they had been represented this way (or with cumulative histograms for each of the 3 weights, as in Figure 9—figure supplement 1)?
c) The text describing Figure 7 says that central brain areas have lower separability indices. Although this is true for OAs vs higher areas overall, in moving from the vestibular nuceli (VN) more centrally, there is a progressive increase in the separability index. The central cells also have more variability in their separability index than the OAs. Some comment about this seems warranted--the existing text provides too superficial consideration of the results.
2) The general message is that "there is a bit of everything everywhere" with only marginal differences across regions; but the complicated analysis (many parameters reported one by one, or 3 at a time Figure 6) and incomplete/indirect reporting (see point 3 below) leaves open another possibility: might there be clusters in the ~ 17 dimension parameter space, which would show a stronger segregation across brain regions?
3) In general, the analysis was well done and well presented, but it might be improved in several ways.
a) This model is "instantaneous", meaning that the rate is related to an instantaneous value of the movement (velocity, acceleration, jerk), and that the history of cell firing is not influencing the firing rate. The three sets of curves (gaussian profile and the 2 first derivative) provide a good set of functions of time for the decomposition of PSTHs: e.g. for a simple profile with a peak, the mean/sd of the peak will correspond to the speed, the assymetry (skewness) will correspond to the acceleration, and the sharpness of the peak (kurtosis) will correspond to the jerk. The shape of the PSTH may indeed be conditioned by the sensitivity of a cell to the speed/acceleration/jerk, but alternate causes, such as long-term impact of movement parameters or of past cellular activity (i.e. violation from the "instantaneous hypothesis"), might also explain the shape of the PSTH; indeed, this is observed in many examples of the Figure 3, where significant departure from baseline are observed at the end of the psth, when virtually all movement is over for hundreds of milliseconds: see Figure 3 (e.g. end of the PSTH Az 315 deg, Elev 45 deg) or Figure 3—figure supplement 1 (e.g. end of Az 270, Elev -45), Figure 3—figure supplement 2 (e.g. Az 90, Elev 0 deg), Figure 3—figure supplement 3 (end of Az 315, Elev 45). All these example suggest that to some degree, the firing of the cells may be affected by past events (movement or cell activity) at much longer time scales than the few tens to hundred milliseconds used for the delay. Therefore, the correspondence between the weights of the fits to a gaussian profile and its first derivatives and an actual sensitivity to velocity, acceleration and jerk is somewhat questionable.
b) A second concern is linked to the use of a non-linear transformation: to describe the spatial dependency of the PSTHs, the sensitivity to the movement components (velocity, acceleration, jerk) is tuned by a non-linear transformation of the cosine of the angle between a (fitted) preferred direction -which may vary across components- and the actual movement direction. The description of these non-linear function is remarkably obfuscated (described by three successive set of variables: a,b,c in y(x)=a.b^x+c, then by y(-1),y(0),y(1), then by Z0 and k0….), but the end result is more clearly shown in Figure 9. One of the important points is that these non-linear functions are not mapping the cosine result (range [-1,1]) onto the full [-1,1] range, but onto a subset of this range. Then, the fit predicts the (spatial dependency of the) rate based on a weighted sum of the non-linearly transformed cosine function. In the MS, theses weights of the components of the fit are used to compare the sensitivities of the cells (e.g. Figure 6). This procedures calls for a few comments:
First, it is unclear if non-linearity is really necessary: inspection of the classes of the 'non-linear functions' (Figure 9) shows set of profiles close to linearity which could be as well approximated by linear (ax+b) functions; the main difference across classes seems more on the target interval than on the shape. The non-linear version will always do better since it permits a finer tuning (1 more parameter), but the benefit may be only marginal (I think the authors could reasonably easily test this) and come at the price a quite obscure process.
Second, because the range over which the cosine is mapped by the non-linear function may be much smaller than the -1,1 range, the weights are not reflecting the two qualitative aspects of the spatial tuning: for example, if a cell has a large gaussian PSTH in all direction, with a 10% modulation of the amplitude of the peak, this will show up as a strong weight of velocity (and a 'broad-positive' non linear function), while it would indeed better described as a strong sensitivity to speed (length of the velocity vector) but a weak sensitivity to velocity (as a vector). In this case, the large weight of the velocity would correspond only to a small dynamic range over spatial orientations; the same large weight for another cell with cosine tuning would correspond to a full dynamic range over spatial orientations. Therefore, a single value of the weight is mixing qualitatively different information (e.g. speed value as in Figure 3—figure supplement 3D, versus velocity vector as in Figure 3—figure supplement 4D). The only things that disambiguate the information are the parameters of the non-linear function, but 1) they are not so trivial (the authors found it is easier to report them as 'arbitrary' categories…), 2) they are reported independently from the values of the weight. A linear version of the transformation ('a.cos(x)+b', provided that it would not degrade too much the performance of the fits) could be easier to report and avoid the categorization: 'a' (in the a.cos(x)+b) would signal the dynamic range over 3D, and 'b' would signal the directionaly-invariant component of the response; it would become possible for example to generate bivariate plots (such as 'weight' vs 'a') which would answer simple questions such as: is the dynamic range for velocity coding identical in all brain areas for strongly velocity-tuned cells? Currently, the choice of reporting makes the answer to this question unclear. Since some of their conclusion is that there are no regional differences, I do not feel so certain they took the most powerful approach to reveal them.
c) It is not clear why the color-coded stacked bar plots are used in Figures 7B-D, and Figure 10C – here the color map is representing PD-angles, which presumably vary continuously, but they have been categorized unnecessarily – cumulative histograms, or at least a continuous color map, here could allow more direct comparisons across regions.
4) The discussion could be more enlightening:
General interest could be broadened by a brief comparison with other sensory systems
Why would the brain maintain such complicated representation of movement (a parameter and its two first derivatives or its integral, with unrelated preferred directions in many cases and non-linear spatial sensitivity to departure from preferred direction) across all brain areas, without further processing beyond the first synapse?
The last paragraph is obscure: what would be the relation between tilt/translation disambiguation and the mixtures of sensitivity described in the study?
Some of the big picture, clearly laid out in the Introduction, is lost in the Results and Discussion sections. For example, it would be useful for the authors to remind us of why we care about space time separabilityhttps://doi.org/10.7554/eLife.20787.024
- Jean Laurens
- Sheng Liu
- Xiong-Jie Yu
- David Dickman
- Dora E Angelaki
- Jean Laurens
- Sheng Liu
- Xiong-Jie Yu
- David Dickman
- Dora E Angelaki
- Gregory C DeAngelis
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
The work was supported by NIH grants EY12814 and DC04260. GCD was supported by EY016178. The authors would like to thank A Chen, Y Gu, A Bryan for recording the central neurons whose properties have been analyzed here.
Animal experimentation: This study was performed in strict accordance with the recommendations in the Guide for the Care and Use of Laboratory Animals of the National Institutes of Health. All of the animals were handled according to approved institutional animal care and use committee (IACUC) protocol #AN-5795 of the Baylor College of Medicine. Data collected in previous studies were obtained according to the National Institutes of Health Guidelines, as described in the corresponding publications. All surgery was performed under isoflurane anesthesia, and every effort was made to minimize suffering.
- Jennifer L Raymond, Reviewing Editor, Stanford University, United States
- Received: August 18, 2016
- Accepted: December 22, 2016
- Version of Record published: January 11, 2017 (version 1)
© 2017, Laurens et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.