Task-dependent optimal representations for cerebellar learning

  1. Marjorie Xie
  2. Samuel P Muscinelli
  3. Kameron Decker Harris
  4. Ashok Litwin-Kumar  Is a corresponding author
  1. Zuckerman Mind Brain Behavior Institute, Columbia University, United States
  2. Department of Computer Science, Western Washington University, United States
7 figures, 1 table and 1 additional file

Figures

Schematic of cerebellar cortex model.

(A) Mossy fiber inputs (blue) project to granule cells (green), which send parallel fibers that contact a Purkinje cell (black). (B) Diagram of neural network model. D task variables are embedded, …

Figure 2 with 4 supplements
Optimal coding level depends on task.

(A) A random categorization task in which inputs are mapped to one of two categories (+1 or –1). Gray plane denotes the decision boundary of a linear classifier separating the two categories. (B) A …

Figure 2—figure supplement 1
Sparse coding levels are sufficient for random categorization tasks irrespective of number of samples, noise level, and dimension.

(A) Error as a function of coding level for networks trained to perform random categorization tasks (as in Figure 2E but with a wider range of associations P). Performance is measured for noisy …

Figure 2—figure supplement 2
Task-dependence of optimal coding level is consistent across activation functions.

Error as a function of coding level for networks with (A) Heaviside and (B) rectified power-law (with power 2, ϕ(u)=max(u,0)2) nonlinearity in the expansion layer. Networks learned Gaussian process targets. …

Figure 2—figure supplement 3
Task-dependence of optimal coding level is consistent across input dimensions.

Error as a function of coding level for networks learning Gaussian process targets with input dimension D=5 (A) and D=7 (B). Dashed lines indicate the performance of a readout of the input layer. …

Figure 2—figure supplement 4
Error as a function of coding level across different values of P and γ.

Dots denote performance of a readout of the expansion layer in simulations. Thin lines denote performance of a readout of the input layer in simulations. Thick lines denote theory for expansion …

Effect of coding level on the expansion layer representation.

(A) Effect of activation threshold on coding level. A point on the surface of the sphere represents a neuron with effective weights Jieff. Blue region represents the set of neurons activated by x, …

Figure 4 with 2 supplements
Frequency decomposition of network and target function.

(A) Geometry of high-dimensional categorization tasks where input patterns are drawn from random, noisy clusters (light regions). Performance depends on overlaps between training patterns from …

Figure 4—figure supplement 1
Error as a function of coding level for learning pure-frequency spherical harmonic functions.

Frequency is indexed by k. Errors are calculated using analytically using Equation 4 and represent the predictions of the theory for an infinitely large expansion. Curves are symmetric around f=0.5

Figure 4—figure supplement 2
Frequency content of categorization tasks.

Power as a function of frequency for random categorization tasks (colors) and for Gaussian process task (black). Power is averaged over realizations of target functions.

Performance of networks with sparse connectivity.

(A) Top: Fully connected network. Bottom: Sparsely connected network with in-degree K<N and excitatory weights with global inhibition onto expansion layer neurons. (B) Error as a function of coding …

Task-independence of optimal anatomical parameters.

(A) Error as a function of in-degree K for networks learning Gaussian process targets. Curves represent different values of γ, the length scale of the Gaussian process. The total number of …

Figure 7 with 2 supplements
Optimal coding level across tasks and neural systems.

(A) Left: Schematic of two-joint arm. Center: Cerebellar cortex model in which sensorimotor task variables at time t are used to predict hand position at time t+δ. Right: Error as a function of …

Figure 7—figure supplement 1
Optimal coding levels in the presence of spiking noise.

(A) Error as a function of coding level in a spiking model. The firing rate of neuron i (in Hz) is given by hiμ=ϕ(gJieffxμθ), where g is a gain term that adjusts the amplitude of the activity and θ is the …

Figure 7—figure supplement 2
Task-dependence of optimal coding level remains consistent under an online climbing fiber-based plasticity rule.

During each epoch of training, the network is presented with all patterns in a randomized order, and the learned weights are updated with each pattern (see Methods). Networks were presented with 30 …

Tables

Table 1
Summary of simulation parameters.

M: number of expansion layer neurons. N: number of input layer neurons. K: number of connections from input layer to a single expansion layer neuron. S: total number of connections from input to …

Figure panelNetwork parametersTask parameters
Figure 2EM=10,000D=50,P=1,000,ϵ=0.1
Figures 2F, 4G and 5B (full)M=200,000D=3,P=30
Figure 5B and EM=200,000,N=7,000,K=4D=3,P=30
Figure 6AS=MK=10,000,N=100,f=0.3D=3,P=200
Figure 6BN=700,K=4,f=0.3D=3,P=200
Figure 6CM=5,000,f=0.3D=3,P=100,γ=1
Figure 6DM=1,000D=3,P=50
Figure 7AM=20,000D=6,P=100; see Methods
Figure 7BM=10,000,N=50,K=7D=50,P=100,ϵ=0.1
Figure 7CM=20,000,N=206,1K3see Methods
Figure 7DM=20,000,N=K=24D=1,P=30; see Methods
Figure 2—figure supplement 1M=10,000See Figure
Figure 2—figure supplement 2M=20,000D=3,P=30
Figure 2—figure supplement 3M=20,000D=3,P=30
Figure 2—figure supplement 4M=20,000D=3
Figure 7—figure supplement 1M=20,000D=3,P=200
Figure 7—figure supplement 2M=10,000,f=0.3D=3,P=30,γ=1

Additional files

Download links