Motor Networks: The Goldilocks zone in neural circuits
Networks of neurons are tough beasts to control. If too many of the connections between the neurons are excitatory, the network becomes hyperactive, driven by feedback from neuron to neuron. But if too many of the connections are inhibitory, the network goes silent, save for a few blips of activity, mercilessly crushed by inhibition. So how do real neuronal networks stay in the Goldilocks zone between too quiet and too loud? How do they maintain stable activity yet remain sensitive to new inputs?
One clue comes from the long-tailed distributions of spiking rates that are seen throughout the cortex: a small number of neurons respond strongly to a specific input, but most spike only weakly, and thus remain ready to spike again in response to a new input (Wohrer et al., 2013; Buzsáki and Mizuseki, 2014). Another clue comes from the balanced network model: according to this model the excitatory and inhibitory inputs to cortical neurons cancel on average, so the neurons can maintain stable, irregular activity (van Vreeswijk and Sompolinsky, 1996; Renart et al., 2010; see Figure 1A,B). However, we don't know how these two clues fit together to explain how neuronal networks reach the Goldilocks zone: in particular, can balanced networks produce long-tailed distributions of spiking rates?
Now, in eLife, Peter Petersen and Rune Berg of the University of Copenhagen report compelling experimental evidence that they can (Petersen and Berg, 2016; Figure 1C). They did this by testing a theoretical model that was published in 2011 (Roxin et al., 2011). The experiments were performed on the spinal network in turtles and combined intracellular and large-scale extracellular recordings of neural activity. The recordings were made during periods of evoked motor behaviour (that is, when the turtle was scratching itself).
Petersen and Berg first showed that the spinal network was balanced. They identified individual neurons in the spinal cord that existed in a regime called the "fluctuation-driven" regime that is characteristic of balanced networks. The membrane voltages of these neurons fluctuated widely between spikes, as expected from neurons receiving the same inhibitory and excitatory input on average (Figure 1B). They also identified neurons that existed in the "mean-driven" regime (Figure 1A): in these neurons the membrane voltages moved from low values to high values rapidly and directly following each spike. Petersen and Berg then showed that it was possible to switch between the fluctuation-driven and mean-driven regimes by changing the balance between the excitatory and inhibitory inputs to the neurons.
Next they tested a slightly off-the-wall prediction made by Alex Roxin and co-workers for neurons in the fluctuation-driven regime (Roxin et al., 2011). This work predicted that if these neurons have an expansive output curve – that is, if the output increases faster than linear as the input increases – then their output spike-rate will have a long-tailed distribution (Figure 1C). Unexpectedly, Petersen and Berg showed that their fluctuation-driven neurons all had such an expansive output curve.
Petersen and Berg then used large-scale population recordings to address the key question: do these expansive output curves give rise to the predicted long-tailed distribution of firing rates across a network? The answer was a resounding yes. The populations they recorded had lognormal distributions of firing rates, and the neurons within each population sat on a continuum between the fluctuation- and mean-driven regimes. Intriguingly, their data suggest that these regimes bore no relation to whether the neurons were inter- or motor-neurons.
Petersen and Berg's work is a rarity in systems neuroscience, an experimental study that tests a computational theory directly, and exhaustively. They have provided compelling evidence that a combination of balanced input and expansive output can hold a network in the Goldilocks zones (that is, keep it both stable and responsive). And by working in the spinal cord networks of the turtle, they were able to show that all these properties exist during ongoing behaviour, and not just during spontaneous neuronal activity. Moreover, they remind us there is nothing privileged about the dynamics of cortical circuits, or the dynamics of neuronal circuits in mammals.
The work also opens up a number of exciting challenges for theory and experiment. Like many behaviours, scratching is a rhythmic action, driven by repeated bursts of spikes. Petersen and Berg focused on the spikes within bursts, but there is silence between bursts. This silence means there must be two timescales for the control of neural activity in the spinal network. Spikes within the bursts arise from fast changes to a neuron's inputs, whether in the fluctuation-driven regime or the mean-driven regime. And the silence between bursts means that the network is able to slowly switch in and out of the driven regimes; that is, it can periodically turn its balanced state on and then off. Clearly we have just started to unpack how neuronal networks control their own activity.
References
-
The log-dynamic brain: how skewed distributions affect network operationsNature Reviews Neuroscience 15:264–278.https://doi.org/10.1038/nrn3687
-
On the distribution of firing rates in networks of cortical neuronsJournal of Neuroscience 31:16217–16226.https://doi.org/10.1523/JNEUROSCI.1677-11.2011
-
Population-wide distributions of neural activity during perceptual decision-makingProgress in Neurobiology 103:156–193.https://doi.org/10.1016/j.pneurobio.2012.09.004
Article and author information
Author details
Publication history
Copyright
© 2016, Humphries
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
-
- 1,264
- views
-
- 145
- downloads
-
- 7
- citations
Views, downloads and citations are aggregated across all versions of this paper published by eLife.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading
-
- Neuroscience
One of the most fundamental laws of physics is the principle of least action. Motivated by its predictive power, we introduce a neuronal least-action principle for cortical processing of sensory streams to produce appropriate behavioral outputs in real time. The principle postulates that the voltage dynamics of cortical pyramidal neurons prospectively minimizes the local somato-dendritic mismatch error within individual neurons. For output neurons, the principle implies minimizing an instantaneous behavioral error. For deep network neurons, it implies the prospective firing to overcome integration delays and correct for possible output errors right in time. The neuron-specific errors are extracted in the apical dendrites of pyramidal neurons through a cortical microcircuit that tries to explain away the feedback from the periphery, and correct the trajectory on the fly. Any motor output is in a moving equilibrium with the sensory input and the motor feedback during the ongoing sensory-motor transform. Online synaptic plasticity reduces the somatodendritic mismatch error within each cortical neuron and performs gradient descent on the output cost at any moment in time. The neuronal least-action principle offers an axiomatic framework to derive local neuronal and synaptic laws for global real-time computation and learning in the brain.
-
- Cell Biology
- Neuroscience
Experience shapes the brain as neural circuits can be modified by neural stimulation or the lack of it. The molecular mechanisms underlying structural circuit plasticity and how plasticity modifies behaviour are poorly understood. Subjective experience requires dopamine, a neuromodulator that assigns a value to stimuli, and it also controls behaviour, including locomotion, learning, and memory. In Drosophila, Toll receptors are ideally placed to translate experience into structural brain change. Toll-6 is expressed in dopaminergic neurons (DANs), raising the intriguing possibility that Toll-6 could regulate structural plasticity in dopaminergic circuits. Drosophila neurotrophin-2 (DNT-2) is the ligand for Toll-6 and Kek-6, but whether it is required for circuit structural plasticity was unknown. Here, we show that DNT-2-expressing neurons connect with DANs, and they modulate each other. Loss of function for DNT-2 or its receptors Toll-6 and kinase-less Trk-like kek-6 caused DAN and synapse loss, impaired dendrite growth and connectivity, decreased synaptic sites, and caused locomotion deficits. In contrast, over-expressed DNT-2 increased DAN cell number, dendrite complexity, and promoted synaptogenesis. Neuronal activity modified DNT-2, increased synaptogenesis in DNT-2-positive neurons and DANs, and over-expression of DNT-2 did too. Altering the levels of DNT-2 or Toll-6 also modified dopamine-dependent behaviours, including locomotion and long-term memory. To conclude, a feedback loop involving dopamine and DNT-2 highlighted the circuits engaged, and DNT-2 with Toll-6 and Kek-6 induced structural plasticity in this circuit modifying brain function and behaviour.