Elsevier

NeuroImage

Volume 159, 1 October 2017, Pages 195-206
NeuroImage

Selective entrainment of brain oscillations drives auditory perceptual organization

https://doi.org/10.1016/j.neuroimage.2017.07.056Get rights and content

Highlights

  • Facing ambiguous soundscapes, humans perceive unitary multistable organizations.

  • Brain oscillations concurrently entrain to the rate of all competing sound patterns.

  • Entrainment to ignored sound organizations is restricted to auditory regions.

  • Entrainment in an auditory-motor network reflects the perceived sound organization.

Abstract

Perceptual sound organization supports our ability to make sense of the complex acoustic environment, to understand speech and to enjoy music. However, the neuronal mechanisms underlying the subjective experience of perceiving univocal auditory patterns that can be listened to, despite hearing all sounds in a scene, are poorly understood. We hereby investigated the manner in which competing sound organizations are simultaneously represented by specific brain activity patterns and the way attention and task demands prime the internal model generating the current percept. Using a selective attention task on ambiguous auditory stimulation coupled with EEG recordings, we found that the phase of low-frequency oscillatory activity dynamically tracks multiple sound organizations concurrently. However, whereas the representation of ignored sound patterns is circumscribed to auditory regions, large-scale oscillatory entrainment in auditory, sensory-motor and executive-control network areas reflects the active perceptual organization, thereby giving rise to the subjective experience of a unitary percept.

Introduction

Perception can be thought of as an act of inference (Gregory, 1980, Helmholtz, 1866). Modern neuroscience views the brain as a predictive machine, continuously generating internal models of the causal dynamics of the world in an attempt to interpret its observations (Bar, 2009, Friston, 2005). Although relevant to all sensory systems, this assumption especially applies to audition (Baldeweg, 2006, Garrido et al., 2009, Winkler et al., 2012). Particularly, it applies to sequential organization, which refers to the sorting of interleaved sounds (Dowling, 1973, Bregman, 1990, Sussman et al., 1999, Shamma et al., 2011, Winkler et al., 2009). Meaningful auditory objects rely on binding distributed spectrotemporal patterns into coherent streams (Bregman, 1990, Nelken and Bar-Yosef, 2009, Sussman et al., 1999). Yet, auditory information can sometimes be feasibly explained by more than one internal model. For instance, in a musical piece, a single note from an instrument could belong simultaneously to a melodic line, to a harmonic progression and to a rhythmic pattern featuring several instruments. However, despite hearing all sounds, we consciously perceive univocal organizations that we can flexibly listen to. Our subjective experience therefore conforms to the Gestalt principle of exclusive allocation (Kohler, 1947), which states that any sensory element should not be used in more than one description of the natural scene at a time. Whether this principle also applies at the neural level, specifying memory representations of the stimulus input (i.e., whether multiple internal models are held simultaneously or only the current attended one) is still a matter of intense debate (Sussman et al., 2014, Denham et al., 2014, Grossberg et al., 2004).

How the brain flexibly assigns individual events to any of the possible perceptual organizations they could fit into is optimally studied with ambiguous, multistable stimulation, because perception depends on the model currently explaining unchanging sensory input (Sterzer et al., 2009). Behavioral evidence on auditory spontaneous perceptual switches suggests that multiple alternative organizations are held simultaneously and compete to describe the acoustic scene (Denham et al., 2014, Pressnitzer and Hupe, 2006, Sterzer et al., 2009, Sussman et al., 2014). Electrophysiological studies in humans have traditionally embedded violations of established regularities within the acoustic streams in order to use change detection auditory evoked potentials, such as the mismatch negativity (MMN) (Näätänen et al., 1978), as an index of sound organization (Sussman et al., 1998, Sussman et al., 1999). However, besides yielding conflicting results, with some studies showing simultaneous encoding of alternative organizations (Pannese et al., 2015, Sussman et al., 2014) while others suggesting that only the currently perceived organization is represented (Sussman et al., 2002, Sussman, 2013, Winkler et al., 2006), evidence of this nature is intrinsically indirect and does not inform about the neural mechanisms underlying the representation of sound organization.

Several studies have shown that any existing regularity in the auditory scene is reflected in oscillatory activity tuned to its temporal pattern (Henry et al., 2014, John et al., 2001, John et al., 2002, Luo et al., 2006, Luo and Poeppel, 2007, Pannese et al., 2015). This is an interesting observation because synchronized oscillatory activity has been proposed as an effective means for neuronal communication (Fries, 2005). Moreover, since the high-excitability phase of ongoing low-frequency oscillations can be selectively entrained to events occurring in an attended stream (Schroeder and Lakatos, 2009), we speculate that neuronal entrainment could underlie our perceptual ability to flexibly reorganize sequential sounds.

We hereby designed a novel ambiguous sound sequence that allowed the study of active perceptual reorganization while controlling for sensory input. Given the quasi-rhythmic nature of most behaviorally relevant acoustic information (Patel, 2008), rhythmic attention (Jones and Boltz, 1989, Large and Jones, 1999), and its neurophysiological counterpart oscillatory entrainment (Herrmann and Henry, 2014, Schroeder and Lakatos, 2009) would likely play a key role (Pannese et al., 2015). Nozaradan et al. (2011) demonstrated that oscillatory entrainment underlies meter imagery, the voluntary organization of musical beats. However, the imagined meter was imposed on a sound sequence with acoustic energy only at the main beat rate. This leaves open the question of whether oscillatory entrainment actually helps to disambiguate a rhythmic structure that has multiple potential meters. With energy at more than one possible meter, task demands may act to enhance the attended meter while suppressing the unattended one, rather than driving the overall meter of the sequence.

To target the dynamics of large-scale neuronal slow oscillatory activity, we combined spectral analyses with source localization of EEG data, seeking to explore the distinction between the neurophysiological nature of simultaneously encoded representations of the auditory scene, and the selected internal model underlying the perceived auditory object.

Section snippets

Participants

Fourteen healthy volunteers (mean age: 28.9 years; age range: 24–38 years; 8 males; 2 left-handed) with no self-reported history of neurological, psychiatric, or hearing impairment and with normal or corrected-to-normal visual acuity participated in the experiment. All participants passed a hearing screening including pure tones of 500, 1000, 2000, and 4000 Hz at 20 dB HL prior to the recording session. One participant reported being an active amateur musician without formal training. Data from

Results

To investigate the neuronal mechanisms underlying the representation and active selection of competing parallel models of sound input, we asked participants to actively listen to an ambiguous sound sequence that could be perceived in two mutually exclusive ways (Fig. 1A; sound1.mp3). The sequence consisted of a melodic ascending-descending pitch pattern with three different tone frequencies separated by one semitone each, at a tone presentation rate of 5 Hz (200 ms SOA) (Fig. 1A). Tone duration

Discussion

The goal of the current study was to investigate neuronal mechanisms underlying the active selection of competing concurrent internal models of the auditory scene, and the neural representations reflecting the sequential organization of the sounds. Our results revealed that multiple competing sound organizations are concurrently represented as entrained oscillatory activity to their intrinsic rhythms. However, we found a clear dissociation between the ignored (non-task-specific) and active

Acknowledgments

This work was supported by the National Institutes of Health (R01 DC004263), the SGR2014-177 grant from the Generalitat de Catalunya, the grant PSI2015-63664-P from MINECO, a FPU grant AP2007-01084 awarded to J.C.F, and the ICREA Acadèmia Distinguished Professorship awarded to C.E. The authors declare no conflict of interests.

References (94)

  • S. Grossberg et al.

    ARTSTREAM: a neural network model of auditory scene analysis and source segregation

    Neural Netw.

    (2004)
  • J. Horvath et al.

    Simultaneously active pre-attentive representations of local and global rules for sound sequences in the human brain

    Cogn. Brain Res.

    (2001)
  • T.P. Jung et al.

    Removal of eye activity artifacts from visual event-related potentials in normal and clinical subjects

    Clin. Neurophysiol.

    (2000)
  • V. Jurcak et al.

    10/20, 10/10, and 10/5 systems revisited: their validity as relative head-surface-based positioning systems

    Neuroimage

    (2007)
  • J. Kayser et al.

    Principal components analysis of Laplacian waveforms as a generic method for identifying ERP generator patterns: II. Adequacy of low-density estimates

    Clin. Neurophysiol.

    (2006)
  • P. Lakatos et al.

    The spectrotemporal filter mechanism of auditory selective attention

    Neuron

    (2013)
  • A.C. Livesey et al.

    Time perception: manipulation of task difficulty dissociates clock functions from other cognitive demands

    Neuropsychologia

    (2007)
  • H. Luo et al.

    Phase patterns of neuronal responses reliably discriminate speech in human auditory cortex

    Neuron

    (2007)
  • E. Maris et al.

    Nonparametric statistical testing of EEG- and MEG-data

    J. Neurosci. Methods

    (2007)
  • B. Morillon et al.

    Predictive motor control of sensory dynamics in auditory active sensing

    Curr. Opin. Neurobiol.

    (2015)
  • R. Näätänen et al.

    Early selective attention effect on evoked potential reinterpreted

    Acta Psychol. (Amst)

    (1978)
  • A. Nobre et al.

    The hazards of time

    Curr. Opin. Neurobiol.

    (2007)
  • D. Pressnitzer et al.

    Temporal dynamics of auditory and visual bistability reveal common principles of perceptual organization

    Curr. Biol.

    (2006)
  • B.H. Repp

    Hearing a melody in different ways: multistability of metrical interpretation, reflected in rate limits of sensorimotor synchronization

    Cognition

    (2007)
  • C.E. Schroeder et al.

    Low-frequency neuronal oscillations as instruments of sensory selection

    Trends Neurosci.

    (2009)
  • C.E. Schroeder et al.

    Dynamics of active sensing and perceptual selection

    Curr. Opin. Neurobiol.

    (2010)
  • S.A. Shamma et al.

    Temporal coherence and attention in auditory scene analysis

    Trends Neurosci.

    (2011)
  • P. Sterzer et al.

    The neural bases of multistable perception

    Trends Cogn. Sci.

    (2009)
  • E. Sussman et al.

    Attention affects the organization of auditory input associated with the mismatch negativity system

    Brain Res.

    (1998)
  • E. Sussman et al.

    Attention effects on auditory scene analysis in children

    Neuropsychologia

    (2009)
  • E. Sussman et al.

    Top-down effects can modify the initially stimulus-driven auditory organization

    Cogn. Brain Res.

    (2002)
  • E.S. Sussman et al.

    Effects of task-switching on neural representations of ambiguous sound input

    Neuropsychologia

    (2014)
  • E.S. Sussman et al.

    Attention modifies sound level detection in young children

    Dev. Cogn. Neurosci.

    (2011)
  • J.R. Tregellas et al.

    Effect of task difficulty on the functional anatomy of temporal processing

    Neuroimage

    (2006)
  • I. Winkler et al.

    Modeling the auditory scene: predictive regularity representations and perceptual objects

    Trends Cogn. Sci.

    (2009)
  • E.M. Zion Golumbic et al.

    Mechanisms underlying selective neuronal tracking of attended speech at a “cocktail party”

    Neuron

    (2013)
  • M. Bar

    The proactive brain: memory for predictions

    Philos. Trans. R. Soc. Lond B Biol. Sci.

    (2009)
  • J. Besle et al.

    Tuning of the human neocortex to the temporal dynamics of attended events

    J. Neurosci.

    (2011)
  • A.S. Bregman

    Auditory Scene Analysis: the Perceptual Organization of Sound. Cambridge (MA): a Bradford Book

    (1990)
  • A.S. Bregman

    Auditory streaming is cumulative

    J. Exp. Psychol. Hum. Percept. Perform.

    (1978)
  • J.L. Chen et al.

    Moving on time: brain network for auditory-motor synchronization is modulated by rhythm complexity and musical training

    J. Cogn. Neurosci.

    (2008)
  • J.L. Chen et al.

    Listening to musical rhythms recruits motor regions of the brain

    Cereb. Cortex

    (2008)
  • J. Costa-Faidella et al.

    Multiple time scales of adaptation in the auditory system as revealed by human evoked potentials

    Psychophysiology

    (2011)
  • T.B. Crapse et al.

    Corollary discharge across the animal kingdom

    Nat. Rev. Neurosci.

    (2008)
  • S. Dehaene et al.

    A neuronal model of a global workspace in effortful cognitive tasks

    Proc. Natl. Acad. Sci. U.S.A.

    (1998)
  • S. Deike et al.

    The build-up of auditory stream segregation: a different perspective

    Front. Psychol.

    (2012)
  • S. Denham et al.

    Stable individual characteristics in the perception of multiple embedded patterns in multistable auditory stimuli

    Front. Neurosci.

    (2014)
  • Cited by (19)

    • Dimension-selective attention and dimensional salience modulate cortical tracking of acoustic dimensions

      2021, NeuroImage
      Citation Excerpt :

      There is some initial evidence that attention may modulate the cortical tracking of different acoustic dimensions. For example, Costa-Faidella et al. (2017) had participants listen to single auditory streams that varied in duration and intensity at different rates; they either responded as to whether five consecutive tones were long or short (duration task) or silently counted the number of loud tones (intensity task). Cortical tracking was stronger for the attended compared to the ignored dimension.

    • Brain oscillations evoked by sound motion

      2021, Brain Research
      Citation Excerpt :

      Induced activity is correlated with experimental event but is not strictly phase-locked to its onset, whereas evoked activity is strictly phase-locked to the onset of an experimental event across trials (Herrmann et al., 2005). This approach has been repeatedly utilized to reveal differences in the degree of phase consistency across different experimental conditions in the studies of selective attention (e.g., Low and Strauss, 2009; Ponjavic-Conte et al., 2013; Costa-Faidella et al., 2017; for reviews, see Engel et al., 2001; Womelsdorf and Fries, 2007; Haegens and Zion Golumbic, 2018; Zoefel et al., 2018). Generally, the phase of slow oscillations may reflect the neuronal mechanisms of sensory processing better than the same oscillations amplitude (VanRullen et al., 2011; Ng et al., 2012).

    • A novel approach to investigate subcortical and cortical sensitivity to temporal structure simultaneously

      2020, Hearing Research
      Citation Excerpt :

      We previously suggested that neural synchronization, and the sustained response, although both reflecting the same stimulus, are perhaps generated in different regions of the brain and are somewhat independent (Herrmann and Johnsrude, 2018a). In the present work, we found that cortical activity synchronized with the click-rate-modulation compared to the unmodulated control stimulus, in line with the results of previous studies (Nozaradan et al., 2011; Costa-Faidella et al., 2017; Henry and Herrmann, 2014). Whether or not the observed neural response at the 3.5 Hz modulation rate reflects entrainment of neural oscillatory activity or a series of evoked potentials cannot not be disentangled in the current study.

    • Auditory streaming and bistability paradigm extended to a dynamic environment

      2019, Hearing Research
      Citation Excerpt :

      Their performance in other tasks (say detecting deviants) could be explored with a time-locked a priori expectation of their perceptual state. Furthermore, perceptual reports in imaging experiments can introduce motor artefacts that could be avoided using a modulated paradigm (see recent papers (Costa-Faidella et al., 2017; Billig et al., 2018; Kondo et al., 2018) and a comprehensive review Snyder and Elhilali (2017)). The same expectation could prove useful for animal models where invasive recording is possible, but objective measures of perception are limited, although see Itatani and Klump (2014), Christison-Lagay and Cohen (2014) and Cai et al. (2018).

    View all citing articles on Scopus
    View full text