[P2 evaluation] Articles

Choisir deux articles dans la liste, provenant de deux intervenants différents (indiqués par leurs initiales). Les articles de BPC ne peuvent être choisis qu'à l'écrit.


AdC1: J Neurosci. 2014 May 7;34(19):6634-9.

Selective modulation of auditory cortical alpha activity in an audiovisual spatial attention task.

Frey JN, Mainy N, Lachaux JP, Müller N, Bertrand O, Weisz N.

Despite substantial research on attentional modulations of visual alpha activity, doubts remain as to the existence and functional relevance of auditory cortical alpha-band oscillations. It has been argued that auditory cortical alpha does not exist, cannot be measured noninvasively, or that it is dependent on visual alpha generators. This study aimed to address these remaining doubts concerning auditory cortical alpha. A magnetoencephalography study was conducted using a combined audiovisual spatial cueing paradigm. In each trial, a cue indicated the side (left or right) and the modality (auditory or visual) to attend, followed by a short lateralized auditory or visual stimulus. Participants were instructed to respond to the stimuli by a button press. Results show that auditory cortical alpha power is selectively modulated by the audiospatial, but not the visuospatial, attention task. These findings provide further evidence for a distinct auditory cortical alpha generator, which can be measured noninvasively.


AdC2: Front Neurosci. 2014 Jun 3;8:132.

How learning to abstract shapes neural sound representations.

Ley A, Vroomen J, Formisano E.

The transformation of acoustic signals into abstract perceptual representations is the essence of the efficient and goal-directed neural processing of sounds in complex natural environments. While the human and animal auditory system is perfectly equipped to process the spectrotemporal sound features, adequate sound identification and categorization require neural sound representations that are invariant to irrelevant stimulus parameters. Crucially, what is relevant and irrelevant is not necessarily intrinsic to the physical stimulus structure but needs to be learned over time, often through integration of information from other senses. This review discusses the main principles underlying categorical sound perception with a special focus on the role of learning and neural plasticity. We examine the role of different neural structures along the auditory processing pathway in the formation of abstract sound representations with respect to hierarchical as well as dynamic and distributed processing models. Whereas most fMRI studies on categorical sound processing employed speech sounds, the emphasis of the current review lies on the contribution of empirical studies using natural or artificial sounds that enable separating acoustic and perceptual processing levels and avoid interference with existing category representations. Finally, we discuss the opportunities of modern analyses techniques such as multivariate pattern analysis (MVPA) in studying categorical sound representations. With their increased sensitivity to distributed activation changes-even in absence of changes in overall signal level-these analyses techniques provide a promising tool to reveal the neural underpinnings of perceptually invariant sound representations.


AdC3: Front Neurosci. 2014 Oct 30;8:348.

Why do I hear but not understand? Stochastic undersampling as a model of degraded neural encoding of speech.

Lopez-Poveda EA.

Hearing impairment is a serious disease with increasing prevalence. It is defined based on increased audiometric thresholds but increased thresholds are only partly responsible for the greater difficulty understanding speech in noisy environments experienced by some older listeners or by hearing-impaired listeners. Identifying the additional factors and mechanisms that impair intelligibility is fundamental to understanding hearing impairment but these factors remain uncertain. Traditionally, these additional factors have been sought in the way the speech spectrum is encoded in the pattern of impaired mechanical cochlear responses. Recent studies, however, are steering the focus toward impaired encoding of the speech waveform in the auditory nerve. In our recent work, we gave evidence that a significant factor might be the loss of afferent auditory nerve fibers, a pathology that comes with aging or noise overexposure. Our approach was based on a signal-processing analogy whereby the auditory nerve may be regarded as a stochastic sampler of the sound waveform and deafferentation may be described in terms of waveform undersampling. We showed that stochastic undersampling simultaneously degrades the encoding of soft and rapid waveform features, and that this degrades speech intelligibility in noise more than in quiet without significant increases in audiometric thresholds. Here, we review our recent work in a broader context and argue that the stochastic undersampling analogy may be extended to study the perceptual consequences of various different hearing pathologies and their treatment.


AdC4: Neuron. 2015 Jul 1;87(1):179-92.

Cortical Membrane Potential Signature of Optimal States for Sensory Signal Detection.

McGinley MJ, David SV, McCormick DA.

The neural correlates of optimal states for signal detection task performance are largely unknown. One hypothesis holds that optimal states exhibit tonically depolarized cortical neurons with enhanced spiking activity, such as occur during movement. We recorded membrane potentials of auditory cortical neurons in mice trained on a challenging tone-in-noise detection task while assessing arousal with simultaneous pupillometry and hippocampal recordings. Arousal measures accurately predicted multiple modes of membrane potential activity, including rhythmic slow oscillations at low arousal, stable hyperpolarization at intermediate arousal, and depolarization during phasic or tonic periods of hyper-arousal. Walking always occurred during hyper-arousal. Optimal signal detection behavior and sound-evoked responses, at both sub-threshold and spiking levels, occurred at intermediate arousal when pre-decision membrane potentials were stably hyperpolarized. These results reveal a cortical physiological signature of the classically observed inverted-U relationship between task performance and arousal and that optimal detection exhibits enhanced sensory-evoked responses and reduced background synaptic activity.


AdC5: Nat Neurosci. 2011 Jun;14(6):775-82.

A normalization model of multisensory integration.

Ohshiro T, Angelaki DE, DeAngelis GC.

Responses of neurons that integrate multiple sensory inputs are traditionally characterized in terms of a set of empirical principles. However, a simple computational framework that accounts for these empirical features of multisensory integration has not been established. We propose that divisive normalization, acting at the stage of multisensory integration, can account for many of the empirical principles of multisensory integration shown by single neurons, such as the principle of inverse effectiveness and the spatial principle. This model, which uses a simple functional operation (normalization) for which there is considerable experimental support, also accounts for the recent observation that the mathematical rule by which multisensory neurons combine their inputs changes with cue reliability. The normalization model, which makes a strong testable prediction regarding cross-modal suppression, may therefore provide a simple unifying computational account of the important features of multisensory integration by neurons.


AdC6: J Neurosci. 2015 Mar 4;35(9):3815-24.

Attending to pitch information inhibits processing of pitch information: the curious case of amusia.

Zendel BR, Lagrois ME, Robitaille N, Peretz I.

In normal listeners, the tonal rules of music guide musical expectancy. In a minority of individuals, known as amusics, the processing of tonality is disordered, which results in severe musical deficits. It has been shown that the tonal rules of music are neurally encoded, but not consciously available in amusics. Previous neurophysiological studies have not explicitly controlled the level of attention in tasks where participants ignored the tonal structure of the stimuli. Here, we test whether access to tonal knowledge can be demonstrated in congenital amusia when attention is controlled. Electric brain responses were recorded while asking participants to detect an individually adjusted near-threshold click in a melody. In half the melodies, a note was inserted that violated the tonal rules of music. In a second task, participants were presented with the same melodies but were required to detect the tonal deviation. Both tasks required sustained attention, thus conscious access to the rules of tonality was manipulated. In the click-detection task, the pitch deviants evoked an early right anterior negativity (ERAN) in both groups. In the pitch-detection task, the pitch deviants evoked an ERAN and P600 in controls but not in amusics. These results indicate that pitch regularities are represented in the cortex of amusics, but are not consciously available. Moreover, performing a pitch-judgment task eliminated the ERAN in amusics, suggesting that attending to pitch information interferes with perception of pitch. We propose that an impaired top-down frontotemporal projection is responsible for this disorder.


AdC7: J Physiol. 2014 Aug 15;592(Pt 16):3371-81.

Hearing in noisy environments: noise invariance and contrast gain control.

Willmore BD, Cooke JE, King AJ.

Contrast gain control has recently been identified as a fundamental property of the auditory system. Electrophysiological recordings in ferrets have shown that neurons continuously adjust their gain (their sensitivity to change in sound level) in response to the contrast of sounds that are heard. At the level of the auditory cortex, these gain changes partly compensate for changes in sound contrast. This means that sounds which are structurally similar, but have different contrasts, have similar neuronal representations in the auditory cortex. As a result, the cortical representation is relatively invariant to stimulus contrast and robust to the presence of noise in the stimulus. In the inferior colliculus (an important subcortical auditory structure), gain changes are less reliably compensatory, suggesting that contrast- and noise-invariant representations are constructed gradually as one ascends the auditory pathway. In addition to noise invariance, contrast gain control provides a variety of computational advantages over static neuronal representations; it makes efficient use of neuronal dynamic range, may contribute to redundancy-reducing, sparse codes for sound and allows for simpler decoding of population responses. The circuits underlying auditory contrast gain control are still under investigation. As in the visual system, these circuits may be modulated by factors other than stimulus contrast, forming a potential neural substrate for mediating the effects of attention as well as interactions between the senses.


AdC8: J Neurosci. 2014 Aug 13;34(33):11152-8.

Microsaccadic responses indicate fast categorization of sounds: a novel approach to study auditory cognition.

Widmann A, Engbert R, Schröger E.

The mental chronometry of the human brain's processing of sounds to be categorized as targets has intensively been studied in cognitive neuroscience. According to current theories, a series of successive stages consisting of the registration, identification, and categorization of the sound has to be completed before participants are able to report the sound as a target by button press after ∼300-500 ms. Here we use miniature eye movements as a tool to study the categorization of a sound as a target or nontarget, indicating that an initial categorization is present already after 80-100 ms. During visual fixation, the rate of microsaccades, the fastest components of miniature eye movements, is transiently modulated after auditory stimulation. In two experiments, we measured microsaccade rates in human participants in an auditory three-tone oddball paradigm (including rare nontarget sounds) and observed a difference in the microsaccade rates between targets and nontargets as early as 142 ms after sound onset. This finding was replicated in a third experiment with directed saccades measured in a paradigm in which tones had to be matched to score-like visual symbols. Considering the delays introduced by (motor) signal transmission and data analysis constraints, the brain must have differentiated target from nontarget sounds as fast as 80-100 ms after sound onset in both paradigms. We suggest that predictive information processing for expected input makes higher cognitive attributes, such as a sound's identity and category, available already during early sensory processing. The measurement of eye movements is thus a promising approach to investigate hearing.


AdC9: Nature. 2015 May 21;521(7552):348-51

Selective corticostriatal plasticity during acquisition of an auditory discrimination task.

Xiong Q, Znamenskiy P, Zador AM.

Perceptual decisions are based on the activity of sensory cortical neurons, but how organisms learn to transform this activity into appropriate actions remains unknown. Projections from the auditory cortex to the auditory striatum carry information that drives decisions in an auditory frequency discrimination task. To assess the role of these projections in learning, we developed a channelrhodopsin-2-based assay to probe selectively for synaptic plasticity associated with corticostriatal neurons representing different frequencies. Here we report that learning this auditory discrimination preferentially potentiates corticostriatal synapses from neurons representing either high or low frequencies, depending on reward contingencies. We observe frequency-dependent corticostriatal potentiation in vivo over the course of training, and in vitro in striatal brain slices. Our findings suggest a model in which the corticostriatal synapses made by neurons tuned to different features of the sound are selectively potentiated to enable the learned transformation of sound into action.


BPC1: Psychophysiology. 2002 Jan;39(1):38-48.

Music matters: preattentive musicality of the human brain.

Koelsch S, Schroger E, Gunter TC.

During listening to a musical piece, unexpected harmonies may evoke brain responses that are reflected electrically as an early right anterior negativity (ERAN) and a late frontal negativity (N5). In the present study we demonstrate that these components of the event-related potential can be evoked preattentively, that is, even when a musical stimulus is ignored. Both ERAN and N5 differed in amplitude as a function of music-theoretical principles. Participants had no special musical expertise; results thus provide evidence for an automatic processing of musical information in nonmusicians.


BPC2: J Neurosci. 2011 Mar 9;31(10):3843-52.

Functional anatomy of language and music perception: temporal and structural factors investigated using functional magnetic resonance imaging.

Rogalsky C, Rong F, Saberi K, Hickok G.

Language and music exhibit similar acoustic and structural properties, and both appear to be uniquely human. Several recent studies suggest that speech and music perception recruit shared computational systems, and a common substrate in Broca's area for hierarchical processing has recently been proposed. However, this claim has not been tested by directly comparing the spatial distribution of activations to speech and music processing within subjects. In the present study, participants listened to sentences, scrambled sentences, and novel melodies. As expected, large swaths of activation for both sentences and melodies were found bilaterally in the superior temporal lobe, overlapping in portions of auditory cortex. However, substantial nonoverlap was also found: sentences elicited more ventrolateral activation, whereas the melodies elicited a more dorsomedial pattern, extending into the parietal lobe. Multivariate pattern classification analyses indicate that even within the regions of blood oxygenation level-dependent response overlap, speech and music elicit distinguishable patterns of activation. Regions involved in processing hierarchical aspects of sentence perception were identified by contrasting sentences with scrambled sentences, revealing a bilateral temporal lobe network. Music perception showed no overlap whatsoever with this network. Broca's area was not robustly activated by any stimulus type. Overall, these findings suggest that basic hierarchical processing for music and speech recruits distinct cortical networks, neither of which involves Broca's area. We suggest that previous claims are based on data from tasks that tap higher-order cognitive processes, such as working memory and/or cognitive control, which can operate in both speech and music domains.


BPC3: Acta Psychol (Amst). 2012 Jul;140(3):230-5.

Shared structural and temporal integration resources for music and arithmetic processing.

Hoch L, Tillmann B.

While previous research has investigated the relationship either between language and music processing or between language and arithmetic processing, the present study investigated the relationship between music and arithmetic processing. Rule-governed number series, with the final number being a correct or incorrect series ending, were visually presented in synchrony with musical sequences, with the final chord functioning as the expected tonic or the less-expected subdominant chord (i.e., tonal function manipulation). Participants were asked to judge the correctness of the final number as quickly and accurately as possible. The results revealed an interaction between the processing of series ending and the processing of the task-irrelevant chords' tonal function, thus suggesting that music and arithmetic processing share cognitive resources. These findings are discussed in terms of general temporal and structural integration resources for linguistic and non-linguistic rule-governed sequences.


BPC4: Cognitive Psychology. 1979. 11, 346-374.

The Psychological Representation of Musical Pitch in a Tonal Context

Krumhansl, CL.

In this series of experiments, evidence was found for a complex psychological representation of musical pitch. The results of a scaling study, in which subjects judged the similarities between pairs of tones presented in an explicitly tonal context, suggest that musical listeners extract a pattern of relationships among tones that is determined not only by pitch height and chroma, but also by member- ship in the major triad chord and the diatonic scale associated with the tonal system of the context. Multidimensional scaling of the similarity ratings gave a three-dimensional conical structure around which the tones were ordered ac- cording to pitch height. The major triad components formed a closely related cluster near the vertex of the cone; the remaining diatonic scale tones formed a less closely related subset farther from the vertex; and, the nondiatonic tones, still farther from the vertex, were widely dispersed. The results also suggest that, in the psychological representation, tones less closely related to the tonality are less stable than tones closely related to the tonality, and that the representation incor- porates the tendency for unstable tones to move toward the more stable tones in time, reflecting the dynamic character of musical tones. In the similarity ratings of the scaling study, tones less related to the tonality were judged more similar to tones more related to the tonality than the reverse temporal order. Furthermore, in a delayed recognition task memory performance for nondiatonic tones was less accurate than for diatonic tones, and nondiatonic tones were more often confused with diatonic tones than diatonic tones were confused with nondiatonic tones. These results indicate the tonality-specific nature of the psychological representa- tion and argue that the perception of music depends not only on psychoacoustic properties of the tones, but also on processes that relate the tones to one another through contact with a well-defined and complex psychological representation of musical pitch.


BPC5: Cereb Cortex. 2009 Mar;19(3):712-23.

Musical training influences linguistic abilities in 8-year-old children: more evidence for brain plasticity.

Moreno S, Marques C, Santos A, Santos M, Castro SL, Besson M.

We conducted a longitudinal study with 32 nonmusician children over 9 months to determine 1) whether functional differences between musician and nonmusician children reflect specific predispositions for music or result from musical training and 2) whether musical training improves nonmusical brain functions such as reading and linguistic pitch processing. Event-related brain potentials were recorded while 8-year-old children performed tasks designed to test the hypothesis that musical training improves pitch processing not only in music but also in speech. Following the first testing sessions nonmusician children were pseudorandomly assigned to music or to painting training for 6 months and were tested again after training using the same tests. After musical (but not painting) training, children showed enhanced reading and pitch discrimination abilities in speech. Remarkably, 6 months of musical training thus suffices to significantly improve behavior and to influence the development of neural processes as reflected in specific pattern of brain waves. These results reveal positive transfer from music to speech and highlight the influence of musical training. Finally, they demonstrate brain plasticity in showing that relatively short periods of training have strong consequences on the functional organization of the children's brain.


BPC6: Neuropsychol Rehabil. 2014;24(6):894-917.

Learning sung lyrics aids retention in normal ageing and Alzheimer's disease.

Moussard A, Bigand E, Belleville S, Peretz I.

Previous studies have suggested that presenting to-be-memorised lyrics in a singing mode, instead of a speaking mode, may facilitate learning and retention in normal adults. In this study, seven healthy older adults and eight participants with mild Alzheimer's disease (AD) learned and memorised lyrics that were either sung or spoken. We measured the percentage of words recalled from these lyrics immediately and after 10 minutes. Moreover, in AD participants, we tested the effect of successive learning episodes for one spoken and one sung excerpt, as well as long-term retention after a four week delay. Sung conditions did not influence lyrics recall in immediate recall but increased delayed recall for both groups. In AD, learning slopes for sung and spoken lyrics did not show a significant difference across successive learning episodes. However, sung lyrics showed a slight advantage over spoken ones after a four week delay. These results suggest that singing may increase the load of initial learning but improve long-term retention of newly acquired verbal information. We further propose some recommendations on how to maximise these effects and make them relevant for therapeutic applications.


BPC7: Science. 1996 Dec 13;274(5294):1926-8.

Statistical learning by 8-month-old infants.

Saffran JR, Aslin RN, Newport EL.

Learners rely on a combination of experience-independent and experience-dependent mechanisms to extract information from the environment. Language acquisition involves both types of mechanisms, but most theorists emphasize the relative importance of experience-independent mechanisms. The present study shows that a fundamental task of language acquisition, segmentation of words from fluent speech, can be accomplished by 8-month-old infants based solely on the statistical relationships between neighboring speech sounds. Moreover, this word segmentation was based on statistical learning from only 2 minutes of exposure, suggesting that infants have access to a powerful mechanism for the computation of statistical properties of the language input.


BPC8: Cognition. 1999 Feb 1;70(1):27-52.

Statistical learning of tone sequences by human infants and adults.

Saffran JR, Johnson EK, Aslin RN, Newport EL.

Previous research suggests that language learners can detect and use the statistical properties of syllable sequences to discover words in continuous speech (e.g. Aslin, R.N., Saffran, J.R., Newport, E.L., 1998. Computation of conditional probability statistics by 8-month-old infants. Psychological Science 9, 321-324; Saffran, J.R., Aslin, R.N., Newport, E.L., 1996. Statistical learning by 8-month-old infants. Science 274, 1926-1928; Saffran, J., R., Newport, E.L., Aslin, R.N., (1996). Word segmentation: the role of distributional cues. Journal of Memory and Language 35, 606-621; Saffran, J.R., Newport, E.L., Aslin, R.N., Tunick, R.A., Barrueco, S., 1997. Incidental language learning: Listening (and learning) out of the corner of your ear. Psychological Science 8, 101-195). In the present research, we asked whether this statistical learning ability is uniquely tied to linguistic materials. Subjects were exposed to continuous non-linguistic auditory sequences whose elements were organized into 'tone words'. As in our previous studies, statistical information was the only word boundary cue available to learners. Both adults and 8-month-old infants succeeded at segmenting the tone stream, with performance indistinguishable from that obtained with syllable streams. These results suggest that a learning mechanism previously shown to be involved in word segmentation can also be used to segment sequences of non-linguistic stimuli.


BPC9: Psychol Sci. 2004 Aug;15(8):511-4.

Music lessons enhance IQ.

Schellenberg EG.

The idea that music makes you smarter has received considerable attention from scholars and the media. The present report is the first to test this hypothesis directly with random assignment of a large sample of children (N = 144) to two different types of music lessons (keyboard or voice) or to control groups that received drama lessons or no lessons. IQ was measured before and after the lessons. Compared with children in the control groups, children in the music groups exhibited greater increases in full-scale IQ. The effect was relatively small, but it generalized across IQ subtests, index scores, and a standardized measure of academic achievement. Unexpectedly, children in the drama group exhibited substantial pre- to post-test improvements in adaptive social behavior that were not evident in the music groups.


BPC10: Neuroimage. 2006 Jul 15;31(4):1771-82.

Cognitive priming in sung and instrumental music: activation of inferior frontal cortex.

Tillmann B, Koelsch S, Escoffier N, Bigand E, Lalitte P, Friederici AD, von Cramon DY.

Neural correlates of the processing of musical syntax-like structures have been investigated via expectancy violation due to musically unrelated (i.e., unexpected) events in musical contexts. Previous studies reported the implication of inferior frontal cortex in musical structure processing. However - due to the strong musical manipulations - activations might be explained by sensory deviance detection or repetition priming. Our present study investigated neural correlates of musical structure processing with subtle musical violations in a musical priming paradigm. Instrumental and sung sequences ended on related and less-related musical targets. The material controlled sensory priming components, and differences in target processing required listeners' knowledge on musical structures. Participants were scanned with functional Magnetic Resonance Imaging (fMRI) while performing speeded phoneme and timbre identification judgments on the targets. Behavioral results acquired in the scanner replicated the facilitation effect of related over less-related targets. The blood oxygen level-dependent (BOLD) signal linked to target processing revealed activation of right inferior frontal areas (i.e., inferior frontal gyrus, frontal operculum, anterior insula) that was stronger for less-related than for related targets, and this was independent of the material carrying the musical structures. This outcome points to the implication of inferior frontal cortex in the processing of syntactic relations also for musical material and to its role in the processing and integration of sequential information over time. In addition to inferior frontal activation, increased activation was observed in orbital gyrus, temporal areas (anterior superior temporal gyrus, posterior superior temporal gyrus and sulcus, posterior middle temporal gyrus) and supramarginal gyrus.


CL1: J Acoust Soc Am. 2008 Dec;124(6):3841-9

Age-related differences in the temporal modulation transfer function with pure-tone carriers.

He NJ, Mills JH, Ahlstrom JB, Dubno JR.

Detection of amplitude modulation (AM) in 500 and 4000 Hz tonal carriers was measured as a function of modulation frequency from younger and older adults with normal hearing through 4000 Hz. The modulation frequency above which sensitivity to AM increased ("transition frequency") was similar for both groups. Temporal modulation transfer function shapes showed significant age-related differences. For younger subjects, AM detection thresholds were generally constant for low modulation frequencies. For a higher carrier frequency, AM detection thresholds then increased as modulation frequency further increased until the transition frequency. In contrast, AM detection for older subjects continuously increased with increasing modulation frequency, indicating an age-related decline in temporal resolution for faster envelope fluctuations. Significant age-related differences were observed whenever AM detection was dependent on temporal cues. For modulation frequencies above the transition frequency, age-related differences were larger for the lower frequency carrier (where both temporal and spectral cues were available) than for the higher frequency carrier (where AM detection was primarily dependent on spectral cues). These results are consistent with a general age-related decline in the synchronization of neural responses to both the carrier waveform and envelope fluctuation.


CL2: J Acoust Soc Am. 2014 Jan;135(1):342-51.

The effects of age and hearing loss on interaural phase difference discrimination.

King A, Hopkins K, Plack CJ

The discrimination of interaural phase differences (IPDs) requires accurate binaural temporal processing and has been used as a measure of sensitivity to temporal envelope and temporal fine structure (TFS). Previous studies found that TFS-IPD discrimination declined with age and with sensorineural hearing loss (SNHL), but age and SNHL have often been confounded. The aim of this study was to determine the independent contributions of age and SNHL to TFS and envelope IPD discrimination by using a sample of adults with a wide range of ages and SNHL. A two-interval, two-alternative forced-choice procedure was used to measure IPD discrimination thresholds for 20-Hz amplitude-modulated tones with carrier frequencies of 250 or 500 Hz when the IPD was in either the stimulus envelope or TFS. There were positive correlations between absolute thresholds and TFS-IPD thresholds, but not envelope-IPD thresholds, when age was accounted for. This supports the idea that SNHL affects TFS processing independently to age. Age was positively correlated with envelope-IPD thresholds at both carrier frequencies and TFS-IPD thresholds at 500 Hz, when absolute thresholds were accounted for. These results suggest that age negatively affects the binaural processing of envelope and TFS at some frequencies independently of SNHL.


CL3: J Assoc Res Otolaryngol. 2013 Oct;14(5):757-66.

Subcortical neural synchrony and absolute thresholds predict frequency discrimination independently.

Marmel F, Linley D, Carlyon RP, Gockel HE, Hopkins K, Plack CJ.

The neural mechanisms of pitch coding have been debated for more than a century. The two main mechanisms are coding based on the profiles of neural firing rates across auditory nerve fibers with different characteristic frequencies (place-rate coding), and coding based on the phase-locked temporal pattern of neural firing (temporal coding). Phase locking precision can be partly assessed by recording the frequency-following response (FFR), a scalp-recorded electrophysiological response that reflects synchronous activity in subcortical neurons. Although features of the FFR have been widely used as indices of pitch coding acuity, only a handful of studies have directly investigated the relation between the FFR and behavioral pitch judgments. Furthermore, the contribution of degraded neural synchrony (as indexed by the FFR) to the pitch perception impairments of older listeners and those with hearing loss is not well known. Here, the relation between the FFR and pure-tone frequency discrimination was investigated in listeners with a wide range of ages and absolute thresholds, to assess the respective contributions of subcortical neural synchrony and other age-related and hearing loss-related mechanisms to frequency discrimination performance. FFR measures of neural synchrony and absolute thresholds independently contributed to frequency discrimination performance. Age alone, i.e., once the effect of subcortical neural synchrony measures or absolute thresholds had been partialed out, did not contribute to frequency discrimination. Overall, the results suggest that frequency discrimination of pure tones may depend both on phase locking precision and on separate mechanisms affected in hearing loss.


CL4: Trends Hear. 2014 Sep 9;18. pii: 2331216514550621

Perceptual consequences of "hidden" hearing loss.

Plack CJ, Barker D, Prendergast G

Dramatic results from recent animal experiments show that noise exposure can cause a selective loss of high-threshold auditory nerve fibers without affecting absolute sensitivity permanently. This cochlear neuropathy has been described as hidden hearing loss, as it is not thought to be detectable using standard measures of audiometric threshold. It is possible that hidden hearing loss is a common condition in humans and may underlie some of the perceptual deficits experienced by people with clinically normal hearing. There is some evidence that a history of noise exposure is associated with difficulties in speech discrimination and temporal processing, even in the absence of any audiometric loss. There is also evidence that the tinnitus experienced by listeners with clinically normal hearing is associated with cochlear neuropathy, as measured using Wave I of the auditory brainstem response. To date, however, there has been no direct link made between noise exposure, cochlear neuropathy, and perceptual difficulties. Animal experiments also reveal that the aging process itself, in the absence of significant noise exposure, is associated with loss of auditory nerve fibers. Evidence from human temporal bone studies and auditory brainstem response measures suggests that this form of hidden loss is common in humans and may have perceptual consequences, in particular, regarding the coding of the temporal aspects of sounds. Hidden hearing loss is potentially a major health issue, and investigations are ongoing to identify the causes and consequences of this troubling condition.


CL5: Hear Res. 2014 Nov;317:50-62.

Amplitude-modulation detection by recreational-noise-exposed humans with near-normal hearing thresholds and its medium-term progression.

Stone MA, Moore BC.

Noise exposure can affect the functioning of cochlear inner and outer hair cells (IHC/OHC), leading to multiple perceptual changes. This work explored possible changes in detection of amplitude modulation (AM) at three Sensation Levels (SL) for carrier frequencies of 3, 4 and 6 kHz. There were two groups of participants, aged 19 to 24 (Young) and 26 to 35 (Older) years. All had near-normal audiometric thresholds. Participants self-assessed exposure to high-level noise in recreational settings. Each group was sub-grouped into low-noise (LN) or high-noise (HN) exposure. AM detection thresholds were worse for the HN than for the LN sub-group at the lowest SL, for the males only of the Young group and for both genders for the Older group, despite no significant difference in absolute threshold at 3 and 4 kHz between sub-groups. AM detection at the lowest SL, at both 3 and 4 kHz, generally improved with increasing age and increasing absolute threshold, consistent with a recruitment-like process. However, poorer AM detection was correlated with increasing exposure at 3 kHz in the Older group. It is suggested that high-level noise exposure produces both IHC- and OHC-related damage, the balance between the two varying across frequency. However, the use of AM detection offers poor sensitivity as a measure of the effects.


CL6: Int J Audiol. 2007 Jan;46(1):39-46.

Ten(HL)-test results and psychophysical tuning curves for subjects with auditory neuropathy.

Vinay, Moore BC

Auditory neuropathy is a hearing disorder characterized by abnormal or absent auditory brainstem responses, and the presence of otoacoustic emissions and/or cochlear microphonics, indicating normal functioning of the outer hair cells. Here, subjects with auditory neuropathy, with near-normal hearing to moderate hearing loss, were tested using the TEN(HL) test for diagnosis of dead regions and also using psychophysical tuning curves (PTCs). Results for the majority of subjects met the TEN(HL)-test criteria at one or more frequencies (often at several or all frequencies). However, the PTCs did not show shifted tips. Hence, the positive results of the TEN(HL) test should not be interpreted as indicating the presence of dead regions. Rather, it appears that high thresholds in noise are caused by poor processing efficiency, perhaps associated with loss of neural synchrony.


DP1: J Acoust Soc Am. 2010 May;127(5):3026-37.

Stimulus uncertainty and insensitivity to pitch-change direction.

Mathias SR, Micheyl C, Bailey PJ.

n a series of experiments, Semal and Demany [(2006). J. Acoust. Soc. Am. 120, 3907-3915] demonstrated that some normally hearing listeners are unable to determine the direction of small but detectable differences in frequency between pure tones. Unlike studies demonstrating similar effects in patients with brain damage, the authors used stimuli in which the standard frequency of the tones was highly uncertain (roved) over trials. In Experiment 1, listeners were identified as insensitive to the direction of pitch changes using stimuli with frequency roving. When listeners were retested using stimuli without roving in Experiment 2, impairments in pitch-direction identification were generally much less profound. In Experiment 3, frequency-roving range had a systematic effect on listeners' thresholds, and impairments in pitch-direction identification tended to occur only when the roving range was widest. In Experiment 4, the influence of frequency roving was similar for continuous frequency changes as for discrete changes. Possible explanations for the influence of roving on listeners' insensitivity to pitch-change direction are discussed.


DP2: J Acoust Soc Am. 2014 Dec;136(6):3350.

Computational speech segregation based on an auditory-inspired modulation analysis.

May T, Dau T.

A monaural speech segregation system is presented that estimates the ideal binary mask from noisy speech based on the supervised learning of amplitude modulation spectrogram (AMS) features. Instead of using linearly scaled modulation filters with constant absolute bandwidth, an auditory-inspired modulation filterbank with logarithmically scaled filters is employed. To reduce the dependency of the AMS features on the overall background noise level, a feature normalization stage is applied. In addition, a spectro-temporal integration stage is incorporated in order to exploit the context information about speech activity present in neighboring time-frequency units. In order to evaluate the generalization performance of the system to unseen acoustic conditions, the speech segregation system is trained with a limited set of low signal-to-noise ratio (SNR) conditions, but tested over a wide range of SNRs up to 20 dB. A systematic evaluation of the system demonstrates that auditory-inspired modulation processing can substantially improve the mask estimation accuracy in the presence of stationary and fluctuating interferers.


DP3: J Acoust Soc Am. 2003 Sep;114(3):1543-9.

Informational masking and musical training.

Oxenham AJ, Fligor BJ, Mason CR, Kidd G Jr.

The relationship between musical training and informational masking was studied for 24 young adult listeners with normal hearing. The listeners were divided into two groups based on musical training. In one group, the listeners had little or no musical training; the other group was comprised of highly trained, currently active musicians. The hypothesis was that musicians may be less susceptible to informational masking, which is thought to reflect central, rather than peripheral, limitations on the processing of sound. Masked thresholds were measured in two conditions, similar to those used by Kidd et al. [J. Acoust. Soc. Am. 95, 3475-3480 (1994)]. In both conditions the signal was comprised of a series of repeated tone bursts at 1 kHz. The masker was comprised of a series of multitone bursts, gated with the signal. In one condition the frequencies of the masker were selected randomly for each burst; in the other condition the masker frequencies were selected randomly for the first burst of each interval and then remained constant throughout the interval. The difference in thresholds between the two conditions was taken as a measure of informational masking. Frequency selectivity, using the notched-noise method, was also estimated in the two groups. The results showed no difference in frequency selectivity between the two groups, but showed a large and significant difference in the amount of informational masking between musically trained and untrained listeners. This informational masking task, which requires no knowledge specific to musical training (such as note or interval names) and is generally not susceptible to systematic short- or medium-term training effects, may provide a basis for further studies of analytic listening abilities in different populations.


DP4: PLoS Comput Biol. 2012 Oct;8(10):e1002731.

How recent history affects perception: the normative approach and its heuristic approximation.

Raviv O, Ahissar M, Loewenstein Y.

There is accumulating evidence that prior knowledge about expectations plays an important role in perception. The Bayesian framework is the standard computational approach to explain how prior knowledge about the distribution of expected stimuli is incorporated with noisy observations in order to improve performance. However, it is unclear what information about the prior distribution is acquired by the perceptual system over short periods of time and how this information is utilized in the process of perceptual decision making. Here we address this question using a simple two-tone discrimination task. We find that the "contraction bias", in which small magnitudes are overestimated and large magnitudes are underestimated, dominates the pattern of responses of human participants. This contraction bias is consistent with the Bayesian hypothesis in which the true prior information is available to the decision-maker. However, a trial-by-trial analysis of the pattern of responses reveals that the contribution of most recent trials to performance is overweighted compared with the predictions of a standard Bayesian model. Moreover, we study participants' performance in a-typical distributions of stimuli and demonstrate substantial deviations from the ideal Bayesian detector, suggesting that the brain utilizes a heuristic approximation of the Bayesian inference. We propose a biologically plausible model, in which decision in the two-tone discrimination task is based on a comparison between the second tone and an exponentially-decaying average of the first tone and past tones. We show that this model accounts for both the contraction bias and the deviations from the ideal Bayesian detector hypothesis. These findings demonstrate the power of Bayesian-like heuristics in the brain, as well as their limitations in their failure to fully adapt to novel environments.


DP5: Curr Opin Neurobiol. 2014 Feb;24(1):103-10.

Single neuron and population coding of natural sounds in auditory cortex.

Mizrahi A, Shalev A, Nelken I.

The auditory system drives behavior using information extracted from sounds. Early in the auditory hierarchy, circuits are highly specialized for detecting basic sound features. However, already at the level of the auditory cortex the functional organization of the circuits and the underlying coding principles become different. Here, we review some recent progress in our understanding of single neuron and population coding in primary auditory cortex, focusing on natural sounds. We discuss possible mechanisms explaining why single neuron responses to simple sounds cannot predict responses to natural stimuli. We describe recent work suggesting that structural features like local subnetworks rather than smoothly mapped tonotopy are essential components of population coding. Finally, we suggest a synthesis of how single neurons and subnetworks may be involved in coding natural sounds.


DP6: Sci Rep. 2015 Sep 24;5:14489

How musical expertise shapes speech perception: evidence from auditory classification images.

Varnet L, Wang T, Peter C, Meunier F, Hoen M.

It is now well established that extensive musical training percolates to higher levels of cognition, such as speech processing. However, the lack of a precise technique to investigate the specific listening strategy involved in speech comprehension has made it difficult to determine how musicians' higher performance in non-speech tasks contributes to their enhanced speech comprehension. The recently developed Auditory Classification Image approach reveals the precise time-frequency regions used by participants when performing phonemic categorizations in noise. Here we used this technique on 19 non-musicians and 19 professional musicians. We found that both groups used very similar listening strategies, but the musicians relied more heavily on the two main acoustic cues, at the first formant onset and at the onsets of the second and third formants onsets. Additionally, they responded more consistently to stimuli. These observations provide a direct visualization of auditory plasticity resulting from extensive musical training and shed light on the level of functional transfer between auditory processing and speech perception.


DP7: Proc Natl Acad Sci U S A. 2014 Jun 24;111(25):E2606-15.

Immersive audiomotor game play enhances neural and perceptual salience of weak signals in noise.

Whitton JP, Hancock KE, Polley DB.

All sensory systems face the fundamental challenge of encoding weak signals in noisy backgrounds. Although discrimination abilities can improve with practice, these benefits rarely generalize to untrained stimulus dimensions. Inspired by recent findings that action video game training can impart a broader spectrum of benefits than traditional perceptual learning paradigms, we trained adult humans and mice in an immersive audio game that challenged them to forage for hidden auditory targets in a 2D soundscape. Both species learned to modulate their angular search vectors and target approach velocities based on real-time changes in the level of a weak tone embedded in broadband noise. In humans, mastery of this tone in noise task generalized to an improved ability to comprehend spoken sentences in speech babble noise. Neural plasticity in the auditory cortex of trained mice supported improved decoding of low-intensity sounds at the training frequency and an enhanced resistance to interference from background masking noise. These findings highlight the potential to improve the neural and perceptual salience of degraded sensory stimuli through immersive computerized games.


MC1: J Neurosci. 2012 Sep 26;32(39):13389-95.

Repetition suppression and expectation suppression are dissociable in time in early auditory evoked fields.

Todorovic A, de Lange FP.

Repetition of a stimulus, as well as valid expectation that a stimulus will occur, both attenuate the neural response to it. These effects, repetition suppression and expectation suppression, are typically confounded in paradigms in which the nonrepeated stimulus is also relatively rare (e.g., in oddball blocks of mismatch negativity paradigms, or in repetition suppression paradigms with multiple repetitions before an alternation). However, recent hierarchical models of sensory processing inspire the hypothesis that the two might be separable in time, with repetition suppression occurring earlier, as a consequence of local transition probabilities, and suppression by expectation occurring later, as a consequence of learnt statistical regularities. Here we test this hypothesis in an auditory experiment by orthogonally manipulating stimulus repetition and stimulus expectation and, using magnetoencephalography, measuring the neural response over time in human subjects. We found that stimulus repetition (but not stimulus expectation) attenuates the early auditory response (40-60 ms), while stimulus expectation (but not stimulus repetition) attenuates the subsequent, intermediate stage of auditory processing (100-200 ms). These findings are well in line with hierarchical predictive coding models, which posit sequential stages of prediction error resolution, contingent on the level at which the hypothesis is generated.


MC2: Curr Biol. 2015 Aug 3;25(15):2051-6.

Human screams occupy a privileged niche in the communication soundscape.

Arnal LH, Flinker A, Kleinschmidt A, Giraud AL, Poeppel D.

Screaming is arguably one of the most relevant communication signals for survival in humans. Despite their practical relevance and their theoretical significance as innate [1] and virtually universal [2, 3] vocalizations, what makes screams a unique signal and how they are processed is not known. Here, we use acoustic analyses, psychophysical experiments, and neuroimaging to isolate those features that confer to screams their alarming nature, and we track their processing in the human brain. Using the modulation power spectrum (MPS [4, 5]), a recently developed, neurally informed characterization of sounds, we demonstrate that human screams cluster within restricted portion of the acoustic space (between ∼30 and 150 Hz modulation rates) that corresponds to a well-known perceptual attribute, roughness. In contrast to the received view that roughness is irrelevant for communication [6], our data reveal that the acoustic space occupied by the rough vocal regime is segregated from other signals, including speech, a pre-requisite to avoid false alarms in normal vocal communication. We show that roughness is present in natural alarm signals as well as in artificial alarms and that the presence of roughness in sounds boosts their detection in various tasks. Using fMRI, we show that acoustic roughness engages subcortical structures critical to rapidly appraise danger. Altogether, these data demonstrate that screams occupy a privileged acoustic niche that, being separated from other communication signals, ensures their biological and ultimately social efficiency.


MC3: Nat Neurosci. 2015 Dec 7.

Cortical tracking of hierarchical linguistic structures in connected speech.

Ding N, Melloni L, Zhang H, Tian X, Poeppel D.

The most critical attribute of human language is its unbounded combinatorial nature: smaller elements can be combined into larger structures on the basis of a grammatical system, resulting in a hierarchy of linguistic units, such as words, phrases and sentences. Mentally parsing and representing such structures, however, poses challenges for speech comprehension. In speech, hierarchical linguistic structures do not have boundaries that are clearly defined by acoustic cues and must therefore be internally and incrementally constructed during comprehension. We found that, during listening to connected speech, cortical activity of different timescales concurrently tracked the time course of abstract linguistic structures at different hierarchical levels, such as words, phrases and sentences. Notably, the neural tracking of hierarchical linguistic structures was dissociated from the encoding of acoustic cues and from the predictability of incoming words. Our results indicate that a hierarchy of neural processing timescales underlies grammar-based internal construction of hierarchical linguistic structure.


MC4: Psychol Sci. 2013 Oct;24(10):1995-2004.

Swinging at a cocktail party: voice familiarity aids speech perception in the presence of a competing voice.

Johnsrude IS, Mackey A, Hakyemez H, Alexander E, Trang HP, Carlyon RP.

People often have to listen to someone speak in the presence of competing voices. Much is known about the acoustic cues used to overcome this challenge, but almost nothing is known about the utility of cues derived from experience with particular voices--cues that may be particularly important for older people and others with impaired hearing. Here, we use a version of the coordinate-response-measure procedure to show that people can exploit knowledge of a highly familiar voice (their spouse's) not only to track it better in the presence of an interfering stranger's voice, but also, crucially, to ignore it so as to comprehend a stranger's voice more effectively. Although performance declines with increasing age when the target voice is novel, there is no decline when the target voice belongs to the listener's spouse. This finding indicates that older listeners can exploit their familiarity with a speaker's voice to mitigate the effects of sensory and cognitive decline.


MC5: Curr Biol. 2015 Aug 31;25(17):2238-46.

Attentive Tracking of Sound Sources.

Woods KJ, McDermott JH.

Auditory scenes often contain concurrent sound sources, but listeners are typically interested in just one of these and must somehow select it for further processing. One challenge is that real-world sounds such as speech vary over time and as a consequence often cannot be separated or selected based on particular values of their features (e.g., high pitch). Here we show that human listeners can circumvent this challenge by tracking sounds with a movable focus of attention. We synthesized pairs of voices that changed in pitch and timbre over random, intertwined trajectories, lacking distinguishing features or linguistic information. Listeners were cued beforehand to attend to one of the voices. We measured their ability to extract this cued voice from the mixture by subsequently presenting the ending portion of one voice and asking whether it came from the cued voice. We found that listeners could perform this task but that performance was mediated by attention-listeners who performed best were also more sensitive to perturbations in the cued voice than in the uncued voice. Moreover, the task was impossible if the source trajectories did not maintain sufficient separation in feature space. The results suggest a locus of attention that can follow a sound's trajectory through a feature space, likely aiding selection and segregation amid similar distractors.


YB1: Neuron. 2012 Oct 18;76(2):435-49.

Discrete neocortical dynamics predict behavioral categorization of sounds.

Bathellier B, Ushakova L, Rumpel S.

The ability to group stimuli into perceptual categories is essential for efficient interaction with the environment. Discrete dynamics that emerge in brain networks are believed to be the neuronal correlate of category formation. Observations of such dynamics have recently been made; however, it is still unresolved if they actually match perceptual categories. Using in vivo two-photon calcium imaging in the auditory cortex of mice, we show that local network activity evoked by sounds is constrained to few response modes. Transitions between response modes are characterized by an abrupt switch, indicating attractor-like, discrete dynamics. Moreover, we show that local cortical responses quantitatively predict discrimination performance and spontaneous categorization of sounds in behaving mice. Our results therefore demonstrate that local nonlinear dynamics in the auditory cortex generate spontaneous sound categories which can be selected for behavioral or perceptual decisions.


YB2: J Neurosci. 2004 Nov 17;24(46):10440-53.

Multiple time scales of adaptation in auditory cortex neurons.

Ulanovsky N, Las L, Farkas D, Nelken I.

Neurons in primary auditory cortex (A1) of cats show strong stimulus-specific adaptation (SSA). In probabilistic settings, in which one stimulus is common and another is rare, responses to common sounds adapt more strongly than responses to rare sounds. This SSA could be a correlate of auditory sensory memory at the level of single A1 neurons. Here we studied adaptation in A1 neurons, using three different probabilistic designs. We showed that SSA has several time scales concurrently, spanning many orders of magnitude, from hundreds of milliseconds to tens of seconds. Similar time scales are known for the auditory memory span of humans, as measured both psychophysically and using evoked potentials. A simple model, with linear dependence on both short-term and long-term stimulus history, provided a good fit to A1 responses. Auditory thalamus neurons did not show SSA, and their responses were poorly fitted by the same model. In addition, SSA increased the proportion of failures in the responses of A1 neurons to the adapting stimulus. Finally, SSA caused a bias in the neuronal responses to unbiased stimuli, enhancing the responses to eccentric stimuli. Therefore, we propose that a major function of SSA in A1 neurons is to encode auditory sensory memory on multiple time scales. This SSA might play a role in stream segregation and in binding of auditory objects over many time scales, a property that is crucial for processing of natural auditory scenes in cats and of speech and music in humans.


YB3: Nat Neurosci. 2003 Apr;6(4):391-8.

Processing of low-probability sounds by cortical neurons.

Ulanovsky N, Las L, Nelken I.

The ability to detect rare auditory events can be critical for survival. We report here that neurons in cat primary auditory cortex (A1) responded more strongly to a rarely presented sound than to the same sound when it was common. For the rare stimuli, we used both frequency and amplitude deviants. Moreover, some A1 neurons showed hyperacuity for frequency deviants--a frequency resolution one order of magnitude better than receptive field widths in A1. In contrast, auditory thalamic neurons were insensitive to the probability of frequency deviants. These phenomena resulted from stimulus-specific adaptation in A1, which may be a single-neuron correlate of an extensively studied cortical potential--mismatch negativity--that is evoked by rare sounds. Our results thus indicate that A1 neurons, in addition to processing the acoustic features of sounds, may also be involved in sensory memory and novelty detection.


YB4: Neuron. 2013 Nov 20;80(4):1066-76.

Optogenetic activation of an inhibitory network enhances feedforward functional connectivity in auditory cortex.

Hamilton LS1, Sohl-Dickstein J, Huth AG, Carels VM, Deisseroth K, Bao S.

The mammalian neocortex is a highly interconnected network of different types of neurons organized into both layers and columns. Overlaid on this structural organization is a pattern of functional connectivity that can be rapidly and flexibly altered during behavior. Parvalbumin-positive (PV+) inhibitory neurons, which are implicated in cortical oscillations and can change neuronal selectivity, may play a pivotal role in these dynamic changes. We found that optogenetic activation of PV+ neurons in the auditory cortex enhanced feedforward functional connectivity in the putative thalamorecipient circuit and in cortical columnar circuits. In contrast, stimulation of PV+ neurons induced no change in connectivity between sites in the same layers. The activity of PV+ neurons may thus serve as a gating mechanism to enhance feedforward, but not lateral or feedback, information flow in cortical circuits. Functionally, it may preferentially enhance the contribution of bottom-up sensory inputs to perception.


YB5: Nat Neurosci. 2004 Sep;7(9):974-81.

Temporal plasticity in the primary auditory cortex induced by operant perceptual learning.

Bao S, Chang EF, Woods J, Merzenich MM.

Processing of rapidly successive acoustic stimuli can be markedly improved by sensory training. To investigate the cortical mechanisms underlying such temporal plasticity, we trained rats in a 'sound maze' in which navigation using only auditory cues led to a target location paired with food reward. In this task, the repetition rate of noise pulses increased as the distance between the rat and target location decreased. After training in the sound maze, neurons in the primary auditory cortex (A1) showed greater responses to high-rate noise pulses and stronger phase-locking of responses to the stimuli; they also showed shorter post-stimulation suppression and stronger rebound activation. These improved temporal dynamics transferred to trains of pure-tone pips. Control animals that received identical sound stimulation but were given free access to food showed the same results as naive rats. We conclude that this auditory perceptual learning results in improvements in temporal processing, which may be mediated by enhanced cortical response dynamics.


YB6: Nat Neurosci. 2011 Jan;14(1):108-14

Auditory cortex spatial sensitivity sharpens during task performance.

Lee CC, Middlebrooks JC.

Activity in the primary auditory cortex (A1) is essential for normal sound localization behavior, but previous studies of the spatial sensitivity of neurons in A1 have found broad spatial tuning. We tested the hypothesis that spatial tuning sharpens when an animal engages in an auditory task. Cats performed a task that required evaluation of the locations of sounds and one that required active listening, but in which sound location was irrelevant. Some 26-44% of the units recorded in A1 showed substantially sharpened spatial tuning during the behavioral tasks as compared with idle conditions, with the greatest sharpening occurring during the location-relevant task. Spatial sharpening occurred on a scale of tens of seconds and could be replicated multiple times in ∼1.5-h test sessions. Sharpening resulted primarily from increased suppression of responses to sounds at least-preferred locations. That and an observed increase in latencies suggest an important role of inhibitory mechanisms.


YB7: J Neurosci. 2012 Feb 29;32(9):3193-210

Activity related to perceptual judgment and action in primary auditory cortex.

Niwa M, Johnson JS, O'Connor KN, Sutter ML.

Recent evidence is reshaping the view of primary auditory cortex (A1) from a unisensory area to one more involved in dynamically integrating multisensory- and task-related information. We found A1 single- (SU) and multiple-unit (MU) activity correlated with macaques' choices in an amplitude modulation (AM) discrimination task. Animals were trained to discriminate AM noise from unmodulated noise by releasing a lever for AM noise and holding down the lever for unmodulated noise. Activity for identical stimuli was compared between trials where the animals reported AM and trials where they did not. We found 47.4% of MUs and 22.8% of SUs significantly increased firing shortly before the animal's behavioral response to report AM when compared to the equivalent time period on trials where AM was not reported. Activity was also linked to lever release in a different task context, suggesting A1 modulation by somatosensory, or efference copy, input. When spikes were counted only during the stimulus, 19.6% of MUs and 13.8% of SUs increased firing rate when animals reported AM compared to when they did not, suggesting an attentional effect, or that A1 activity can be used by higher decision areas, or that such areas provide feedback to A1. Activity associated with AM reporting was correlated with a unit's AM sensitivity, suggesting AM sensitive neurons' involvement in task performance. A1 neurons' phase locking to AM correlated more weakly (compared to firing rate) with the animals' report of AM, suggesting a preferential role for rate-codes in A1 for this AM discrimination task.


YB8: Nat Neurosci. 2009 May;12(5):646-54.

Engaging in an auditory task suppresses responses in auditory cortex.

Otazu GH, Tai LH, Yang Y, Zador AM.

Although systems that are involved in attentional selection have been studied extensively, much less is known about nonselective systems. To study these preparatory mechanisms, we compared activity in auditory cortex that was elicited by sounds while rats performed an auditory task ('engaged') with activity that was elicited by identical stimuli while subjects were awake but not performing a task ('passive'). We found that engagement suppressed responses, an effect that was opposite in sign to that elicited by selective attention. In the auditory thalamus, however, engagement enhanced spontaneous firing rates but did not affect evoked responses. These results indicate that neural activity in auditory cortex cannot be viewed simply as a limited resource that is allocated in greater measure as the state of the animal passes from somnolent to passively listening to engaged and attentive. Instead, the engaged condition possesses a characteristic and distinct neural signature in which sound-evoked responses are paradoxically suppressed.


YB9: Neuron. 2014 Jun 4;82(5):1157-70.

Neural correlates of task switching in prefrontal cortex and primary auditory cortex in a novel stimulus selection task for rodents.

Rodgers CC, DeWeese MR.

Animals can selectively respond to a target sound despite simultaneous distractors, just as humans can respond to one voice at a crowded cocktail party. To investigate the underlying neural mechanisms, we recorded single-unit activity in primary auditory cortex (A1) and medial prefrontal cortex (mPFC) of rats selectively responding to a target sound from a mixture. We found that prestimulus activity in mPFC encoded the selection rule-which sound from the mixture the rat should select. Moreover, electrically disrupting mPFC significantly impaired performance. Surprisingly, prestimulus activity in A1 also encoded selection rule, a cognitive variable typically considered the domain of prefrontal regions. Prestimulus changes correlated with stimulus-evoked changes, but stimulus tuning was not strongly affected. We suggest a model in which anticipatory activation of a specific network of neurons underlies the selection of a sound from a mixture, giving rise to robust and widespread rule encoding in both brain regions.


YB10: J Neurosci. 2011 Aug 17;31(33):11867-78.

Extra-classical tuning predicts stimulus-dependent receptive fields in auditory neurons.

Schneider DM, Woolley SM.

The receptive fields of many sensory neurons are sensitive to statistical differences among classes of complex stimuli. For example, excitatory spectral bandwidths of midbrain auditory neurons and the spatial extent of cortical visual neurons differ during the processing of natural stimuli compared to the processing of artificial stimuli. Experimentally characterizing neuronal nonlinearities that contribute to stimulus-dependent receptive fields is important for understanding how neurons respond to different stimulus classes in multiple sensory modalities. Here we show that in the zebra finch, many auditory midbrain neurons have extra-classical receptive fields, consisting of sideband excitation and sideband inhibition. We also show that the presence, degree, and asymmetry of stimulus-dependent receptive fields during the processing of complex sounds are predicted by the presence, valence, and asymmetry of extra-classical tuning. Neurons for which excitatory bandwidth expands during the processing of song have extra-classical excitation. Neurons for which frequency tuning is static and for which excitatory bandwidth contracts during the processing of song have extra-classical inhibition. Simulation experiments further demonstrate that stimulus-dependent receptive fields can arise from extra-classical tuning with a static spike threshold nonlinearity. These findings demonstrate that a common neuronal nonlinearity can account for the stimulus dependence of receptive fields estimated from the responses of auditory neurons to stimuli with natural and non-natural statistics.


YB11: Nat Neurosci. 2014 Jun;17(6):841-50.

Scaling down of balanced excitation and inhibition by active behavioral states in auditory cortex.

Zhou M, Liang F, Xiong XR, Li L, Li H, Xiao Z, Tao HW, Zhang LI.

Cortical sensory processing is modulated by behavioral and cognitive states. How this modulation is achieved by changing synaptic circuits remains largely unknown. In awake mouse auditory cortex, we found that sensory-evoked spike responses of layer 2/3 (L2/3) excitatory cells were scaled down with preserved sensory tuning when mice transitioned from quiescence to active behaviors, including locomotion, whereas L4 and thalamic responses were unchanged. Whole-cell voltage-clamp recordings revealed that tone-evoked synaptic excitation and inhibition exhibited a robust functional balance. The change to active states caused scaling down of excitation and inhibition at approximately equal levels in L2/3 cells, but resulted in no synaptic changes in L4 cells. This lamina-specific gain control could be attributed to an enhancement of L1-mediated inhibitory tone, with L2/3 parvalbumin inhibitory neurons also being suppressed. Thus, L2/3 circuits can adjust the salience of output in accordance with momentary behavioral demands while maintaining the sensitivity and quality of sensory processing.