1. Neuroscience
Download icon

Motor Networks: The Goldilocks zone in neural circuits

  1. Mark D Humphries  Is a corresponding author
  1. University of Manchester, United Kingdom
  • Cited 3
  • Views 1,189
  • Annotations
Cite this article as: eLife 2016;5:e22735 doi: 10.7554/eLife.22735


How do networks of neurons remain both stable and sensitive to new inputs?

Main text

Networks of neurons are tough beasts to control. If too many of the connections between the neurons are excitatory, the network becomes hyperactive, driven by feedback from neuron to neuron. But if too many of the connections are inhibitory, the network goes silent, save for a few blips of activity, mercilessly crushed by inhibition. So how do real neuronal networks stay in the Goldilocks zone between too quiet and too loud? How do they maintain stable activity yet remain sensitive to new inputs?

One clue comes from the long-tailed distributions of spiking rates that are seen throughout the cortex: a small number of neurons respond strongly to a specific input, but most spike only weakly, and thus remain ready to spike again in response to a new input (Wohrer et al., 2013; Buzsáki and Mizuseki, 2014). Another clue comes from the balanced network model: according to this model the excitatory and inhibitory inputs to cortical neurons cancel on average, so the neurons can maintain stable, irregular activity (van Vreeswijk and Sompolinsky, 1996; Renart et al., 2010; see Figure 1A,B). However, we don't know how these two clues fit together to explain how neuronal networks reach the Goldilocks zone: in particular, can balanced networks produce long-tailed distributions of spiking rates?

Long-tailed distributions and balanced networks.

(A) When a single neuron receives more excitatory input (red) than inhibitory input (blue), its membrane voltage (black line) spikes regularly. In this "mean-driven" model the membrane voltage spends much of its time at or near the spiking threshold (see histogram on right), and spiking is driven by the difference between the means of the excitatory and inhibitory inputs. However, most neurons do not spike as rapidly or regularly as predicted by the "mean-driven" model. (B) The theory of balanced networks proposes that the inhibitory and excitatory inputs to a neuron have approximately the same mean and the same variance. Consequently, spiking is driven by random fluctuations in the inputs, when the excitatory input briefly exceeds the inhibitory input. A signature of this “fluctuation-driven” regime is that the distribution of the membrane voltage is approximately Gaussian (histogram). Fluctuation-driven neurons spike slowly and irregularly, similar to the majority of the neurons in the cortex. (C) The relationship between the input to a neuron (blue curve below the x-axis) and its spiking or firing rate (blue curve to the left of the y-axis) is described by a firing rate versus input (f–i) curve. In 2011 Roxin et al. predicted that an expansive f–i curve (see main text) would convert a Gaussian input into a long-tailed distribution of firing rates in the fluctuation-driven regime. Moreover, according to this theory, if each neuron has a Gaussian input, then the overall population of neurons will also have a Gaussian input, and if each neuron has an expansive f–i curve in the fluctuation-driven regime, then the output of the population will be long-tailed.

Now, in eLife, Peter Petersen and Rune Berg of the University of Copenhagen report compelling experimental evidence that they can (Petersen and Berg, 2016; Figure 1C). They did this by testing a theoretical model that was published in 2011 (Roxin et al., 2011). The experiments were performed on the spinal network in turtles and combined intracellular and large-scale extracellular recordings of neural activity. The recordings were made during periods of evoked motor behaviour (that is, when the turtle was scratching itself).

Petersen and Berg first showed that the spinal network was balanced. They identified individual neurons in the spinal cord that existed in a regime called the "fluctuation-driven" regime that is characteristic of balanced networks. The membrane voltages of these neurons fluctuated widely between spikes, as expected from neurons receiving the same inhibitory and excitatory input on average (Figure 1B). They also identified neurons that existed in the "mean-driven" regime (Figure 1A): in these neurons the membrane voltages moved from low values to high values rapidly and directly following each spike. Petersen and Berg then showed that it was possible to switch between the fluctuation-driven and mean-driven regimes by changing the balance between the excitatory and inhibitory inputs to the neurons.

Next they tested a slightly off-the-wall prediction made by Alex Roxin and co-workers for neurons in the fluctuation-driven regime (Roxin et al., 2011). This work predicted that if these neurons have an expansive output curve – that is, if the output increases faster than linear as the input increases – then their output spike-rate will have a long-tailed distribution (Figure 1C). Unexpectedly, Petersen and Berg showed that their fluctuation-driven neurons all had such an expansive output curve.

Petersen and Berg then used large-scale population recordings to address the key question: do these expansive output curves give rise to the predicted long-tailed distribution of firing rates across a network? The answer was a resounding yes. The populations they recorded had lognormal distributions of firing rates, and the neurons within each population sat on a continuum between the fluctuation- and mean-driven regimes. Intriguingly, their data suggest that these regimes bore no relation to whether the neurons were inter- or motor-neurons.

Petersen and Berg's work is a rarity in systems neuroscience, an experimental study that tests a computational theory directly, and exhaustively. They have provided compelling evidence that a combination of balanced input and expansive output can hold a network in the Goldilocks zones (that is, keep it both stable and responsive). And by working in the spinal cord networks of the turtle, they were able to show that all these properties exist during ongoing behaviour, and not just during spontaneous neuronal activity. Moreover, they remind us there is nothing privileged about the dynamics of cortical circuits, or the dynamics of neuronal circuits in mammals.

The work also opens up a number of exciting challenges for theory and experiment. Like many behaviours, scratching is a rhythmic action, driven by repeated bursts of spikes. Petersen and Berg focused on the spikes within bursts, but there is silence between bursts. This silence means there must be two timescales for the control of neural activity in the spinal network. Spikes within the bursts arise from fast changes to a neuron's inputs, whether in the fluctuation-driven regime or the mean-driven regime. And the silence between bursts means that the network is able to slowly switch in and out of the driven regimes; that is, it can periodically turn its balanced state on and then off. Clearly we have just started to unpack how neuronal networks control their own activity.


Article and author information

Author details

  1. Mark D Humphries

    Faculty of Biology, Medicine and Health, University of Manchester, Manchester, United Kingdom
    For correspondence
    Competing interests
    The author declares that no competing interests exist.
    ORCID icon "This ORCID iD identifies the author of this article:" 0000-0002-1906-2581

Publication history

  1. Version of Record published: December 2, 2016 (version 1)


© 2016, Humphries

This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.


  • 1,189
    Page views
  • 143
  • 3

Article citation count generated by polling the highest count across the following sources: Crossref, PubMed Central, Scopus.

Download links

A two-part list of links to download the article, or parts of the article, in various formats.

Downloads (link to download the article as PDF)

Download citations (links to download the citations from this article in formats compatible with various reference manager tools)

Open citations (links to open the citations from this article in various online reference manager services)

Further reading

    1. Neuroscience
    Yunzhe Liu et al.
    Research Article Updated

    There are rich structures in off-task neural activity which are hypothesized to reflect fundamental computations across a broad spectrum of cognitive functions. Here, we develop an analysis toolkit – temporal delayed linear modelling (TDLM) – for analysing such activity. TDLM is a domain-general method for finding neural sequences that respect a pre-specified transition graph. It combines nonlinear classification and linear temporal modelling to test for statistical regularities in sequences of task-related reactivations. TDLM is developed on the non-invasive neuroimaging data and is designed to take care of confounds and maximize sequence detection ability. Notably, as a linear framework, TDLM can be easily extended, without loss of generality, to capture rodent replay in electrophysiology, including in continuous spaces, as well as addressing second-order inference questions, for example, its temporal and spatial varying pattern. We hope TDLM will advance a deeper understanding of neural computation and promote a richer convergence between animal and human neuroscience.

    1. Neuroscience
    Cornelius Schröder et al.
    Research Article Updated

    Many sensory systems use ribbon-type synapses to transmit their signals to downstream circuits. The properties of this synaptic transfer fundamentally dictate which aspects in the original stimulus will be accentuated or suppressed, thereby partially defining the detection limits of the circuit. Accordingly, sensory neurons have evolved a wide variety of ribbon geometries and vesicle pool properties to best support their diverse functional requirements. However, the need for diverse synaptic functions does not only arise across neuron types, but also within. Here we show that UV-cones, a single type of photoreceptor of the larval zebrafish eye, exhibit striking differences in their synaptic ultrastructure and consequent calcium to glutamate transfer function depending on their location in the eye. We arrive at this conclusion by combining serial section electron microscopy and simultaneous ‘dual-colour’ two-photon imaging of calcium and glutamate signals from the same synapse in vivo. We further use the functional dataset to fit a cascade-like model of the ribbon synapse with different vesicle pool sizes, transfer rates, and other synaptic properties. Exploiting recent developments in simulation-based inference, we obtain full posterior estimates for the parameters and compare these across different retinal regions. The model enables us to extrapolate to new stimuli and to systematically investigate different response behaviours of various ribbon configurations. We also provide an interactive, easy-to-use version of this model as an online tool. Overall, we show that already on the synaptic level of single-neuron types there exist highly specialised mechanisms which are advantageous for the encoding of different visual features.