Transformation of a temporal speech cue to a spatial neural code in human auditory cortex

  1. Neal P Fox
  2. Matthew Leonard
  3. Matthias J Sjerps
  4. Edward F Chang  Is a corresponding author
  1. University of California, San Francisco, United States
  2. Donders Institute for Brain, Cognition and Behaviour, Radboud University, Netherlands

Abstract

In speech, listeners extract continuously-varying spectrotemporal cues from the acoustic signal to perceive discrete phonetic categories. Spectral cues are spatially encoded in the amplitude of responses in phonetically-tuned neural populations in auditory cortex. It remains unknown whether similar neurophysiological mechanisms encode temporal cues like voice-onset time (VOT), which distinguishes sounds like /b/-/p/. We used direct brain recordings in humans to investigate the neural encoding of temporal speech cues with a VOT continuum from /ba/ to /pa/. We found that distinct neural populations respond preferentially to VOTs from one phonetic category, and are also sensitive to sub-phonetic VOT differences within a population's preferred category. In a simple neural network model, simulated populations tuned to detect either temporal gaps or coincidences between spectral cues captured encoding patterns observed in real neural data. These results demonstrate that a spatial/amplitude neural code underlies the cortical representation of both spectral and temporal speech cues.

Data availability

Data and code are available under a Creative Commons License at the project page on Open Science Framework (https://osf.io/9y7uh/).

The following data sets were generated

Article and author information

Author details

  1. Neal P Fox

    Neurological Surgery, University of California, San Francisco, San Francisco, United States
    Competing interests
    The authors declare that no competing interests exist.
    ORCID icon "This ORCID iD identifies the author of this article:" 0000-0003-0298-3664
  2. Matthew Leonard

    Department of Neurological Surgery, University of California, San Francisco, San Francisco, United States
    Competing interests
    The authors declare that no competing interests exist.
  3. Matthias J Sjerps

    Centre for Cognitive Neuroimaging, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, Netherlands
    Competing interests
    The authors declare that no competing interests exist.
  4. Edward F Chang

    Department of Neurological Surgery, University of California, San Francisco, San Francisco, United States
    For correspondence
    edward.chang@ucsf.edu
    Competing interests
    The authors declare that no competing interests exist.
    ORCID icon "This ORCID iD identifies the author of this article:" 0000-0003-2480-4700

Funding

National Institutes of Health (R01-DC012379)

  • Edward F Chang

National Institutes of Health (F32-DC015966)

  • Neal P Fox

European Commission (FP7-623072)

  • Matthias J Sjerps

New York Stem Cell Foundation

  • Edward F Chang

William K. Bowes, Jr. Foundation

  • Edward F Chang

Howard Hughes Medical Institute

  • Edward F Chang

Shurl and Kay Curci Foundation

  • Edward F Chang

The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.

Ethics

Human subjects: All participants gave their written informed consent before surgery and affirmed it at the start of each recording session. The study protocol was approved by the University of California, San Francisco Committee on Human Research. (Protocol number 10-03842: Task-evoked changes in the electrocorticogram in epilepsy patients undergoing invasive electrocorticography and cortical mapping for the surgical treatment of intractable seizures)

Copyright

© 2020, Fox et al.

This article is distributed under the terms of the Creative Commons Attribution License permitting unrestricted use and redistribution provided that the original author and source are credited.

Metrics

  • 1,884
    views
  • 315
    downloads
  • 18
    citations

Views, downloads and citations are aggregated across all versions of this paper published by eLife.

Download links

A two-part list of links to download the article, or parts of the article, in various formats.

Downloads (link to download the article as PDF)

Open citations (links to open the citations from this article in various online reference manager services)

Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)

  1. Neal P Fox
  2. Matthew Leonard
  3. Matthias J Sjerps
  4. Edward F Chang
(2020)
Transformation of a temporal speech cue to a spatial neural code in human auditory cortex
eLife 9:e53051.
https://doi.org/10.7554/eLife.53051

Share this article

https://doi.org/10.7554/eLife.53051

Further reading

    1. Neuroscience
    Proloy Das, Mingjian He, Patrick L Purdon
    Tools and Resources

    Modern neurophysiological recordings are performed using multichannel sensor arrays that are able to record activity in an increasingly high number of channels numbering in the 100s to 1000s. Often, underlying lower-dimensional patterns of activity are responsible for the observed dynamics, but these representations are difficult to reliably identify using existing methods that attempt to summarize multivariate relationships in a post hoc manner from univariate analyses or using current blind source separation methods. While such methods can reveal appealing patterns of activity, determining the number of components to include, assessing their statistical significance, and interpreting them requires extensive manual intervention and subjective judgment in practice. These difficulties with component selection and interpretation occur in large part because these methods lack a generative model for the underlying spatio-temporal dynamics. Here, we describe a novel component analysis method anchored by a generative model where each source is described by a bio-physically inspired state-space representation. The parameters governing this representation readily capture the oscillatory temporal dynamics of the components, so we refer to it as oscillation component analysis. These parameters – the oscillatory properties, the component mixing weights at the sensors, and the number of oscillations – all are inferred in a data-driven fashion within a Bayesian framework employing an instance of the expectation maximization algorithm. We analyze high-dimensional electroencephalography and magnetoencephalography recordings from human studies to illustrate the potential utility of this method for neuroscience data.

    1. Neuroscience
    Sihan Yang, Anastasia Kiyonaga
    Insight

    A neural signature of serial dependence has been found, which mirrors the attractive bias of visual information seen in behavioral experiments.