[P2 evaluation] Oral

Choisir un seul article dans la liste. La liste est désormais définitive, les articles des cours ADC et BT sont en fin de page.


ALG1 : Neuron. 2003 Jun 5;38(5):831-42.

Dissociating neural mechanisms of temporal sequencing and processing phonemes.

Gelfand JR, Bookheimer SY.

Using fMRI, we sought to determine whether the posterior, superior portion of Broca's area performs operations on phoneme segments specifically or implements processes general to sequencing discrete units. Twelve healthy volunteers performed two sequence manipulation tasks and one matching task, using strings of syllables and hummed notes. The posterior portion of Broca's area responded specifically to the sequence manipulation tasks, independent of whether the stimuli were composed of phonemes or hummed notes. In contrast, the left supramarginal gyrus was somewhat more specific to sequencing phoneme segments. These results suggest a functional dissociation of the canonical left hemisphere language regions encompassing the "phonological loop," with the left posterior inferior frontal gyrus responding not to the sound structure of language but rather to sequential operations that may underlie the ability to form words out of dissociable elements.


ALG2 : Neuron. 2005 Nov 23;48(4):687-97

Human dorsal and ventral auditory streams subserve rehearsal-based and echoic processes during verbal working memory.

Buchsbaum BR, Olsen RK, Koch P, Berman KF.

To hear a sequence of words and repeat them requires sensory-motor processing and something more-temporary storage. We investigated neural mechanisms of verbal memory by using fMRI and a task designed to tease apart perceptually based ("echoic") memory from phonological-articulatory memory. Sets of two- or three-word pairs were presented bimodally, followed by a cue indicating from which modality (auditory or visual) items were to be retrieved and rehearsed over a delay. Although delay-period activation in the planum temporale (PT) was insensible to the source modality and showed sustained delay-period activity, the superior temporal gyrus (STG) activated more vigorously when the retrieved items had arrived to the auditory modality and showed transient delay-period activity. Functional connectivity analysis revealed two topographically distinct fronto-temporal circuits, with STG co-activating more strongly with ventrolateral prefrontal cortex and PT co-activating more strongly with dorsolateral prefrontal cortex. These argue for separate contributions of ventral and dorsal auditory streams in verbal working memory.


ALG3 : J Neurosci. 2003 Apr 15;23(8):3423-31.

Hierarchical processing in spoken language comprehension.

Davis MH, Johnsrude IS.

Understanding spoken language requires a complex series of processing stages to translate speech sounds into meaning. In this study, we use functional magnetic resonance imaging to explore the brain regions that are involved in spoken language comprehension, fractionating this system into sound-based and more abstract higher-level processes. We distorted English sentences in three acoustically different ways, applying each distortion to varying degrees to produce a range of intelligibility (quantified as the number of words that could be reported) and collected whole-brain echo-planar imaging data from 12 listeners using sparse imaging. The blood oxygenation level-dependent signal correlated with intelligibility along the superior and middle temporal gyri in the left hemisphere and in a less-extensive homologous area on the right, the left inferior frontal gyrus (LIFG), and the left hippocampus. Regions surrounding auditory cortex, bilaterally, were sensitive to intelligibility but also showed a differential response to the three forms of distortion, consistent with sound-form-based processes. More distant intelligibility-sensitive regions within the superior and middle temporal gyri, hippocampus, and LIFG were insensitive to the acoustic form of sentences, suggesting more abstract nonacoustic processes. The hierarchical organization suggested by these results is consistent with cognitive models and auditory processing in nonhuman primates. Areas that were particularly active for distorted speech conditions and, thus, might be involved in compensating for distortion, were found exclusively in the left hemisphere and partially overlapped with areas sensitive to intelligibility, perhaps reflecting attentional modulation of auditory and linguistic processes.


ALG4 : Neuron. 2005 Sep 15;47(6):893-905.

High binaural coherence determines successful sound localization and increased activity in posterior auditory areas.

Zimmer U, Macaluso E.

Our brain continuously receives complex combinations of sounds originating from different sources and relating to different events in the external world. Timing differences between the two ears can be used to localize sounds in space, but only when the inputs to the two ears have similar spectrotemporal profiles (high binaural coherence). We used fMRI to investigate any modulation of auditory responses by binaural coherence. We assessed how processing of these cues depends on whether spatial information is task relevant and whether brain activity correlates with subjects' localization performance. We found that activity in Heschl's gyrus increased with increasing coherence, irrespective of whether localization was task relevant. Posterior auditory regions also showed increased activity for high coherence, primarily when sound localization was required and subjects successfully localized sounds. We conclude that binaural coherence cues are processed throughout the auditory cortex and that these cues are used in posterior regions for successful auditory localization.


ALG5 : J Neurophysiol. 2004 Mar;91(3):1282-96.

Tonotopic organization in human auditory cortex revealed by progressions of frequency sensitivity.

Talavage TM, Sereno MI, Melcher JR, Ledden PJ, Rosen BR, Dale AM.

Functional neuroimaging experiments have revealed an organization of frequency-dependent responses in human auditory cortex suggestive of multiple tonotopically organized areas. Numerous studies have sampled cortical responses to isolated narrow-band stimuli, revealing multiple locations in auditory cortex at which the position of response varies systematically with frequency content. Because appropriate anatomical or functional grouping of these distinct frequency-dependent responses is uncertain, the number and location of tonotopic mappings within human auditory cortex remains unclear. Further, sampling does not address whether the observed mappings exhibit continuity as a function of position. This functional magnetic resonance imaging study used frequency-swept stimuli to identify progressions of frequency sensitivity across the cortical surface. The center-frequency of narrow-band, amplitude-modulated noise was slowly swept between 125 and 8,000 Hz. The latency of response relative to sweep onset was determined for each cortical surface location. Because frequency varied systematically with time, response latency indicated the frequency to which a location was maximally sensitive. Areas of cortex exhibiting a progressive change in response latency with position were considered tonotopically organized. There exist two main findings. First, six progressions of frequency sensitivity (i.e., tonotopic mappings) were repeatably observed in the superior temporal plane. Second, the locations of the higher- and lower-frequency endpoints of these progressions were approximately congruent with regions reported to be most responsive to discrete higher- and lower-frequency stimuli. Based on these findings and previous anatomical work, we propose a correspondence between these progressions and anatomically defined cortical areas, suggesting that five areas in human auditory cortex exhibit at least six tonotopic organizations.


CL1 : Ear Hear. 2004 Jun;25(3):242-50.

Temporal fine-structure cues to speech and pure tone modulation in observers with sensorineural hearing loss.

Buss E, Hall JW 3rd, Grose JH.

OBJECTIVE: The purpose of this study was to examine the effect of sensorineural hearing loss on the ability to make use of fine temporal information and to evaluate the relation between this ability and the ability to recognize speech. DESIGN: Fourteen observers with normal hearing and 12 observers with sensorineural hearing loss were tested on open-set word recognition and on psychophysical tasks thought to reflect use of fine-structure cues: the detection of 2 Hz frequency modulation (FM) and the discrimination of the rate of amplitude modulation (AM) and quasifrequency modulation (QFM). RESULTS: The results showed relatively poor performance for observers with sensorineural hearing loss on both the speech recognition and psychoacoustical tasks. Of particular interest was the finding of significant correlations within the hearing-loss group between speech recognition performance and the psychoacoustical tasks based on frequency modulation, which are thought to reflect the quality of the coding of temporal fine structure. CONCLUSIONS: These results suggest that sensorineural hearing loss may be associated with a reduced ability to use fine temporal information that is coded by neural phase-locking to stimulus fine-structure and that this may contribute to poor speech recognition performance and to poor performance on psychoacoustical tasks that depend on temporal fine structure. Copyright 2004 Lippincott Williams and Wilkins


CL2 : J Acoust Soc Am. 1992 May;91(5):2881-93.

Pitch discrimination and phase sensitivity in young and elderly subjects and its relationship to frequency selectivity.

Moore BC, Peters RW.

Frequency difference limens for pure tones (DLFs) and for complex tones (DLCs) were measured for four groups of subjects: young normal hearing, young hearing impaired, elderly with near-normal hearing, and elderly hearing impaired. The auditory filters of the subjects had been measured in earlier experiments using the notched-noise method, for center frequencies (fc) of 100, 200, 400, and 800 Hz. The DLFs for both impaired groups were higher than for the young normal group at all fc's (50-4000 Hz). The DLFs at a given fc were generally only weakly correlated with the sharpness of the auditory filter at that fc, and some subjects with broad filters had near-normal DLFs at low frequencies. Some subjects in the elderly normal group had very large DLFs at low frequencies in spite of near-normal auditory filters. These results suggest a partial dissociation of frequency selectivity and frequency discrimination of pure tones. The DLCs for the two impaired groups were higher than those for the young normal group at all fundamental frequencies (fo) tested (50, 100, 200, and 400 Hz); the DLCs for the elderly normal group were intermediate. At fo = 50 Hz, DLCs for a complex tone containing only low harmonics (1-5) were markedly higher than for complex tones containing higher harmonics, for all subject groups, suggesting that pitch was conveyed largely by the higher, unresolved harmonics. For the elderly impaired group, and some subjects in the elderly normal group, DLCs were larger for a complex tone with lower harmonics (1-12) than for tones without lower harmonics (4-12 and 6-12) for fo's up to 200 Hz. Some elderly normal subjects had markedly larger-than-normal DLCs in spite of near-normal auditory filters. The DLCs tended to be larger for complexes with components added in alternating sine/cosine phase than for complexes with components added in cosine phase. Phase effects were significant for all groups, but were small for the young normal group. The results are not consistent with place-based models of the pitch perception of complex tones; rather, they suggest that pitch is at least partly determined by temporal mechanisms.


CL3 : J Acoust Soc Am. 2003 Feb;113(2):961-8.

Understanding speech in modulated interference: cochlear implant users and normal-hearing listeners.

Nelson PB, Jin SH, Carney AE, Nelson DA.

Many competing noises in real environments are modulated or fluctuating in level. Listeners with normal hearing are able to take advantage of temporal gaps in fluctuating maskers. Listeners with sensorineural hearing loss show less benefit from modulated maskers. Cochlear implant users may be more adversely affected by modulated maskers because of their limited spectral resolution and by their reliance on envelope-based signal-processing strategies of implant processors. The current study evaluated cochlear implant users' ability to understand sentences in the presence of modulated speech-shaped noise. Normal-hearing listeners served as a comparison group. Listeners repeated IEEE sentences in quiet, steady noise, and modulated noise maskers. Maskers were presented at varying signal-to-noise ratios (SNRs) at six modulation rates varying from 1 to 32 Hz. Results suggested that normal-hearing listeners obtain significant release from masking from modulated maskers, especially at 8-Hz masker modulation frequency. In contrast, cochlear implant users experience very little release from masking from modulated maskers. The data suggest, in fact, that they may show negative effects of modulated maskers at syllabic modulation rates (2-4 Hz). Similar patterns of results were obtained from implant listeners using three different devices with different speech-processor strategies. The lack of release from masking occurs in implant listeners independent of their device characteristics, and may be attributable to the nature of implant processing strategies and/or the lack of spectral detail in processed stimuli.


CL4 : Proc Natl Acad Sci U S A. 2005 Feb 15;102(7):2293-8.

Speech recognition with amplitude and frequency modulations.

Zeng FG, Nie K, Stickney GS, Kong YY, Vongphoe M, Bhargave A, Wei C, Cao K.

Amplitude modulation (AM) and frequency modulation (FM) are commonly used in communication, but their relative contributions to speech recognition have not been fully explored. To bridge this gap, we derived slowly varying AM and FM from speech sounds and conducted listening tests using stimuli with different modulations in normal-hearing and cochlear-implant subjects. We found that although AM from a limited number of spectral bands may be sufficient for speech recognition in quiet, FM significantly enhances speech recognition in noise, as well as speaker and tone recognition. Additional speech reception threshold measures revealed that FM is particularly critical for speech recognition with a competing voice and is independent of spectral resolution and similarity. These results suggest that AM and FM provide independent yet complementary contributions to support robust speech recognition under realistic listening situations. Encoding FM may improve auditory scene analysis, cochlear-implant, and audiocoding performance.


DP1 : Nature. 2005 Aug 25;436(7054):1161-5.

The neuronal representation of pitch in primate auditory cortex.

Bendor D, Wang X.

Pitch perception is critical for identifying and segregating auditory objects, especially in the context of music and speech. The perception of pitch is not unique to humans and has been experimentally demonstrated in several animal species. Pitch is the subjective attribute of a sound's fundamental frequency (f(0)) that is determined by both the temporal regularity and average repetition rate of its acoustic waveform. Spectrally dissimilar sounds can have the same pitch if they share a common f(0). Even when the acoustic energy at f(0) is removed ('missing fundamental') the same pitch is still perceived. Despite its importance for hearing, how pitch is represented in the cerebral cortex is unknown. Here we show the existence of neurons in the auditory cortex of marmoset monkeys that respond to both pure tones and missing fundamental harmonic complex sounds with the same f(0), providing a neural correlate for pitch constancy. These pitch-selective neurons are located in a restricted low-frequency cortical region near the anterolateral border of the primary auditory cortex, and is consistent with the location of a pitch-selective area identified in recent imaging studies in humans.


DP2 : J Acoust Soc Am. 1999 Nov;106(5):2805-11.

Memory for pitch versus memory for loudness.

Clement S, Demany L, Semal C.

The decays of pitch traces and loudness traces in short-term auditory memory were compared in forced-choice discrimination experiments. The two stimuli presented on each trial were separated by a variable delay (D); they consisted of pure tones, series of resolved harmonics, or series of unresolved harmonics mixed with lowpass noise. A roving procedure was employed in order to minimize the influence of context coding. During an initial phase of each experiment, frequency and intensity discrimination thresholds [P(C) = 0.80] were measured with an adaptive staircase method while D was fixed at 0.5 s. The corresponding physical differences (in cents or dB) were then constantly presented at four values of D: 0.5, 2, 5, and 10 s. In the case of intensity discrimination, performance (d') markedly decreased when D increased from 0.5 to 2 s, but was not further reduced when D was longer. In the case of frequency discrimination, the decline of performance as a function of D was significantly less abrupt. This divergence suggests that pitch and loudness are processed in separate modules of auditory memory.


DP3 : J Exp Psychol Hum Percept Perform. 2003 Jun;29(3):713-25.

Perceptual asymmetries in audition.

Cusack R, Carlyon RP.

Visual feature extraction has been investigated using search experiments. Targets that contain a feature not present in the distractors are easier to detect than if they do not, leading to search asymmetries. If sounds are decomposed into features in the auditory system, there might be asymmetries in analogous tasks. Six experiments investigating this are described. Strong asymmetries were identified, with frequency-modulated targets easier to detect among pure-tone distractors than vice versa and longer sounds easier to select from short distractors than the reverse. It is demonstrated that this asymmetry is not a result of peripheral limitations. In contrast, no asymmetries were observed between high- and low-frequency tones or between short 3-tone sequences differing only in their temporal structure. The results are discussed with reference to models of perceptual grouping and attention, the applicability of analogies between vision and audition, and possible physiological correlates. The paradigm provides a new way in which to investigate auditory feature extraction.


DP4 : Exp Psychol. 2004;51(4):240-8.

The memory of noise.

Kaernbach C.

The memory of auditory random waveforms (i.e., noise) is a special case of auditory memory for sensory information. Five experiments are reported that evaluate the dynamics of this storage system as well as interactions with new input. Periodic waveforms can be discriminated from uncorrelated noise by naive listeners up to a cycle length of 20 s, with the major decline in performance between 5 and 10 s. Even single repetitions of a piece of the waveform can be detected up to a stimulus onset asynchrony (SOA) of 6 s. The capacity of this storage system is limited to a few items of, in total, a few hundred milliseconds length. Within this capacity, however, items do not interfere strongly. These results are compatible with the view that auditory sensory memory is a modality-specific module of short-term memory.


DP5 : J Neurosci. 2005 Feb 9;25(6):1503-13.

Representation of tone in fluctuating maskers in the ascending auditory system.

Las L, Stern EA, Nelken I.

Humans and animals detect low-level tones masked by slowly fluctuating noise very efficiently. A possible neuronal correlate of this phenomenon is the ability of low-level tones to suppress neuronal locking to the envelope of the fluctuating noise ("locking suppression"). Using in vivo intracellular and extracellular recordings in cats, we studied neuronal responses to combinations of fluctuating noise and tones in three successive auditory stations: inferior colliculus (IC), medial geniculate body (MGB), and primary auditory cortex (A1). We found that although the most sensitive responses in the IC were approximately isomorphic to the physical structure of the sounds, with only a small perturbation in the responses to the fluctuating noise after the addition of low-level tones, some neurons in the MGB and all A1 neurons displayed striking suppressive effects. These neurons were hypersensitive, showing suppression already with tone levels lower than the threshold of the neurons in silence. The hypersensitive locking suppression in A1 and MGB had a special timing structure, starting >75 ms after tone onset. Our findings show a qualitative change in the representation of tone in fluctuating noise along the IC-MGB-A1 axis, suggesting the gradual segregation of signal from noise and the representation of the signal as a separate perceptual object in A1.


DP6 : Percept Psychophys. 1994 Oct;56(4):414-23.

Absolute memory for musical pitch: evidence from the production of learned melodies.

Levitin DJ.

Evidence for the absolute nature of long-term auditory memory is provided by analyzing the production of familiar melodies. Additionally, a two-component theory of absolute pitch is presented, in which this rare ability is conceived as consisting of a more common ability, pitch memory, and a separate, less common ability, pitch labeling. Forty-six subjects sang two different popular songs, and their productions were compared with the actual pitches used in recordings of those songs. Forty percent of the subjects sang the correct pitch on at least one trial; 12% of the subjects hit the correct pitch on both trials, and 44% came within two semitones of the correct pitch on both trials. The results show a convergence with previous studies on the stability of auditory imagery and latent absolute pitch ability; the results further suggest that individuals might possess representations of pitch that are more stable and accurate than previously recognized.


JME1 : J Neurosci. 2005 Mar 9;25(10):2490-503.

Plasticity in primary auditory cortex of monkeys with altered vocal production.

Cheung SW, Nagarajan SS, Schreiner CE, Bedenbaugh PH, Wong A.

Response properties of primary auditory cortical neurons in the adult common marmoset monkey (Callithrix jacchus) were modified by extensive exposure to altered vocalizations that were self-generated and rehearsed frequently. A laryngeal apparatus modification procedure permanently lowered the frequency content of the native twitter call, a complex communication vocalization consisting of a series of frequency modulation (FM) sweeps. Monkeys vocalized shortly after this procedure and maintained voicing efforts until physiological evaluation 5-15 months later. The altered twitter calls improved over time, with FM sweeps approaching but never reaching the normal spectral range. Neurons with characteristic frequencies <4.3 kHz that had been weakly activated by native twitter calls were recruited to encode self-uttered altered twitter vocalizations. These neurons showed a decrease in response magnitude and an increase in temporal dispersion of response timing to twitter call and parametric FM stimuli but a normal response profile to pure tone stimuli. Tonotopic maps in voice-modified monkeys were not distorted. These findings suggest a previously unrecognized form of cortical plasticity that is specific to higher-order processes involved in the discrimination of more complex sounds, such as species-specific vocalizations.


JME2 : Brain Res. 2001 Feb 9;891(1-2):78-93.

In vivo Hebbian and basal forebrain stimulation treatment in morphologically identified auditory cortical cells.

Cruikshank SJ, Weinberger NM.

The present study concerns the interactions of local pre/postsynaptic covariance and activity of the cortically-projecting cholinergic basal forebrain, in physiological plasticity of auditory cortex. Specifically, a tone that activated presynaptic inputs to a recorded auditory cortical neuron was repeatedly paired with a combination of two stimuli: (1) local juxtacellular current that excited the recorded cell and (2) basal forebrain stimulation which desynchronized the cortical EEG. In addition, the recorded neurons were filled with biocytin for morphological examination. The hypothesis tested was that the combined treatment would cause increased potentiation of responses to the paired tone, relative to similar conditioning treatments involving either postsynaptic excitation alone or basal forebrain stimulation alone. In contrast, there was no net increase in plasticity and indeed the combined treatment appears to have decreased plasticity below that previously found for either treatment alone. Several alternate interpretations of these results are discussed.


JME3 : J Neurophysiol. 2004 Jul;92(1):73-82.

Environmental enrichment improves response strength, threshold, selectivity, and latency of auditory cortex neurons.

Engineer ND, Percaccio CR, Pandya PK, Moucha R, Rathbun DL, Kilgard MP.

Over the last 50 yr, environmental enrichment has been shown to generate more than a dozen changes in brain anatomy. The consequences of these physical changes on information processing have not been well studied. In this study, rats were housed in enriched or standard conditions either prior to or after reaching sexual maturity. Evoked potentials from awake rats and extracellular recordings from anesthetized rats were used to document responses of auditory cortex neurons. This report details several significant, new findings about the influence of housing conditions on the responses of rat auditory cortex neurons. First, enrichment dramatically increases the strength of auditory cortex responses. Tone-evoked potentials of enriched rats, for example, were more than twice the amplitude of rats raised in standard laboratory conditions. Second, cortical responses of both young and adult animals benefit from exposure to an enriched environment and are degraded by exposure to an impoverished environment. Third, housing condition resulted in rapid remodeling of cortical responses in <2 wk. Fourth, recordings made under anesthesia indicate that enrichment increases the number of neurons activated by any sound. This finding shows that the evoked potential plasticity documented in awake rats was not due to differences in behavioral state. Finally, enrichment made primary auditory cortex (A1) neurons more sensitive to quiet sounds, more selective for tone frequency, and altered their response latencies. These experiments provide the first evidence of physiologic changes in auditory cortex processing resulting from generalized environmental enrichment.


JME4 : J Neurosci. 2004 Feb 4;24(5):1089-100.

Linearity of cortical receptive fields measured with natural sounds.

Machens CK, Wehr MS, Zador AM.

How do cortical neurons represent the acoustic environment? This question is often addressed by probing with simple stimuli such as clicks or tone pips. Such stimuli have the advantage of yielding easily interpreted answers, but have the disadvantage that they may fail to uncover complex or higher-order neuronal response properties. Here, we adopt an alternative approach, probing neuronal responses with complex acoustic stimuli, including animal vocalizations. We used in vivo whole-cell methods in the rat auditory cortex to record subthreshold membrane potential fluctuations elicited by these stimuli. Most neurons responded robustly and reliably to the complex stimuli in our ensemble. Using regularization techniques, we estimated the linear component, the spectrotemporal receptive field (STRF), of the transformation from the sound (as represented by its time-varying spectrogram) to the membrane potential of the neuron. We find that the STRF has a rich dynamical structure, including excitatory regions positioned in general accord with the prediction of the classical tuning curve. However, whereas the STRF successfully predicts the responses to some of the natural stimuli, it surprisingly fails completely to predict the responses to others; on average, only 11% of the response power could be predicted by the STRF. Therefore, most of the response of the neuron cannot be predicted by the linear component, although the response is deterministically related to the stimulus. Analysis of the systematic errors of the STRF model shows that this failure cannot be attributed to simple nonlinearities such as adaptation to mean intensity, rectification, or saturation. Rather, the highly nonlinear response properties of auditory cortical neurons must be attributable to nonlinear interactions between sound frequencies and time-varying properties of the neural encoder.


JME5 : J Neurosci. 2005 Jan 19;25(3):699-705.

Enriched acoustic environment after noise trauma reduces hearing loss and prevents cortical map reorganization.

Norena AJ, Eggermont JJ.

Exposure to sound of sufficient duration and level causes permanent damage to the peripheral auditory system, which results in the reorganization of the cortical tonotopic map. The changes are such that neurons with pre-exposure tuning to frequencies in the hearing loss range now become tuned to frequencies near the near-normal lower boundary of the hearing loss range, which thus becomes over represented. However, cats exposed to a traumatizing noise and immediately thereafter placed for a few weeks in an enriched acoustic environment presented a much-restricted hearing loss compared with similarly exposed cats that were placed for the same time in a quiet environment. The enriched environment spectrally matched the expected hearing loss range and was approximately 40 dB above the level of the expected hearing loss. The hearing loss in the quiet environment-reared cats ranged from 6 to 32 kHz with the largest loss (on average, 40 dB) ranging from 24 to 32 kHz. In contrast, the hearing loss in the enriched-environment cats was restricted to 6-8 kHz at a level of, on average, 35 dB and with 16-32 kHz having normal thresholds. Despite the remaining hearing loss for the enriched-environment cats in the 6-8 kHz range, plastic tonotopic map changes in primary auditory cortex could no longer be demonstrated, suggesting that the enriched acoustic environment prevents this reorganization. This finding has implications for the treatment of hearing disorders, such as tinnitus, that have been linked to cortical tonotopic map reorganization.


ADC1 : J Acoust Soc Am. 2005 Aug;118(2):977-81.

Can dichotic pitches form two streams?

Akeroyd MA, Carlyon RP, Deeks JM.

The phenomenon of auditory streaming reflects the perceptual organization of sounds over time. A series of "A" and "B" tones, presented in a repeating "ABA-ABA" sequence, may be perceived as one "galloping" stream or as two separate streams, depending on the presentation rate and the A-B frequency separation. The present experiment examined whether streaming occurs for sequences of "Huggins pitches," for which the percepts of pitch are derived from the binaural processing of a sharp transition in interaural phase in an otherwise diotic noise. Ten-second "ABA" sequences were presented to eight normal-hearing listeners for two types of stimuli: Huggins-pitch stimuli with interaural phase transitions centered on frequencies between 400 and 800 Hz, or partially-masked diotic tones-in-noise, acting as controls. Listeners indicated, throughout the sequence, the number of streams perceived. The results showed that, for both Huggins-pitch stimuli and tones-in-noise, two streams were often reported. In both cases, the amount of streaming built up over time, and depended on the frequency separation between the A and B tones. These results provide evidence that streaming can occur between stimuli whose pitch percept is derived binaurally. They are inconsistent with models of streaming based solely on differences in the monaural excitation pattern.


ADC2 : J Acoust Soc Am. 2003 Sep;114(3):1543-9.

Informational masking and musical training.

Oxenham AJ, Fligor BJ, Mason CR, Kidd G Jr.

The relationship between musical training and informational masking was studied for 24 young adult listeners with normal hearing. The listeners were divided into two groups based on musical training. In one group, the listeners had little or no musical training; the other group was comprised of highly trained, currently active musicians. The hypothesis was that musicians may be less susceptible to informational masking, which is thought to reflect central, rather than peripheral, limitations on the processing of sound. Masked thresholds were measured in two conditions, similar to those used by Kidd et al. [J. Acoust. Soc. Am. 95, 3475-3480 (1994)]. In both conditions the signal was comprised of a series of repeated tone bursts at 1 kHz. The masker was comprised of a series of multitone bursts, gated with the signal. In one condition the frequencies of the masker were selected randomly for each burst; in the other condition the masker frequencies were selected randomly for the first burst of each interval and then remained constant throughout the interval. The difference in thresholds between the two conditions was taken as a measure of informational masking. Frequency selectivity, using the notched-noise method, was also estimated in the two groups. The results showed no difference in frequency selectivity between the two groups, but showed a large and significant difference in the amount of informational masking between musically trained and untrained listeners. This informational masking task, which requires no knowledge specific to musical training (such as note or interval names) and is generally not susceptible to systematic short- or medium-term training effects, may provide a basis for further studies of analytic listening abilities in different populations.


ADC3 : J Acoust Soc Am. 1990 Aug;88(2):806-12.

Restarting the adapted binaural system.

Hafter ER, Buell TN.

Previous experiments using trains of high-frequency filtered clicks have shown that for lateralization based on interaural difference of time or level, there is a decline in the usefulness of interaural information after the signal's onset when the clicks are presented at a high rate. This process has been referred to as "binaural adaptation." Of interest here are the conditions that produce a recovery from adaptation and allow for a resampling of the interaural information. A train of clicks with short interclick intervals is used to produce adaptation. Then, during its course, a treatment such as the insertion of a temporal gap or the addition of another "triggering" sound is tested for its ability to restart the binaural process. All of the brief triggers tested are shown to be capable of promoting recovery from adaptation. This suggests that, while the binaural system deals with the demands of high-frequency stimulation with rapid adaptation, it quickly cancels the adaptation in response to stimulus change.


ADC4 : J Acoust Soc Am. 2004 Aug;116(2):1092-104.

Across-frequency interference effects in fundamental frequency discrimination: questioning evidence for two pitch mechanisms.

Gockel H, Carlyon RP, Plack CJ.

Carlyon and Shackleton [J. Acoust. Soc. Am. 95, 3541-3554 (1994)] presented an influential study supporting the existence of two pitch mechanisms, one for complex tones containing resolved and one for complex tones containing only unresolved components. The current experiments provide an alternative explanation for their finding, namely the existence of across-frequency interference in fundamental frequency (F0) discrimination. Sensitivity (d') was measured for F0 discrimination between two sequentially presented 400 ms complex (target) tones containing only unresolved components. In experiment 1, the target was filtered between 1375 and 15,000 Hz, had a nominal F0 of 88 Hz, and was presented either alone or with an additional complex tone ("interferer"). The interferer was filtered between 125-625 Hz, and its F0 varied between 88 and 114.4 Hz across blocks. Sensitivity was significantly reduced in the presence of the interferer, and this effect decreased as its F0 was moved progressively further from that of the target. Experiment 2 showed that increasing the level of a synchronously gated lowpass noise that spectrally overlapped with the interferer reduced this "pitch discrimination interference (PDI)". In experiment 3A, the target was filtered between 3900 and 5400 Hz and had an F0 of either 88 or 250 Hz. It was presented either alone or with an interferer, filtered between 1375 and 1875 Hz with an F0 corresponding to the nominal target F0. PDI was larger in the presence of the resolved (250 Hz F0) than in the presence of the unresolved (88 Hz F0) interferer, presumably because the pitch of the former was more salient than that of the latter. Experiments 4A and 4B showed that PDI was reduced but not eliminated when the interferer was gated on 200 ms before and off 200 ms after the target, and that some PDI was observed with a continuous interferer. The current findings provide an alternative interpretation of a study supposedly providing strong evidence for the existence of two pitch mechanisms.


ADC5 : Proc Natl Acad Sci U S A. 2000 Oct 24;97(22):11773-9.

Detection of synchrony in the activity of auditory nerve fibers by octopus cells of the mammalian cochlear nucleus.

Oertel D, Bal R, Gardner SM, Smith PH, Joris PX.

The anatomical and biophysical specializations of octopus cells allow them to detect the coincident firing of groups of auditory nerve fibers and to convey the precise timing of that coincidence to their targets. Octopus cells occupy a sharply defined region of the most caudal and dorsal part of the mammalian ventral cochlear nucleus. The dendrites of octopus cells cross the bundle of auditory nerve fibers just proximal to where the fibers leave the ventral and enter the dorsal cochlear nucleus, each octopus cell spanning about one-third of the tonotopic array. Octopus cells are excited by auditory nerve fibers through the activation of rapid, calcium-permeable, alpha-amino-3-hydroxy-5-methyl-4-isoxazole-propionate receptors. Synaptic responses are shaped by the unusual biophysical characteristics of octopus cells. Octopus cells have very low input resistances (about 7 M Omega), and short time constants (about 200 microsec) as a consequence of the activation at rest of a hyperpolarization-activated mixed-cation conductance and a low-threshold, depolarization-activated potassium conductance. The low input resistance causes rapid synaptic currents to generate rapid and small synaptic potentials. Summation of small synaptic potentials from many fibers is required to bring an octopus cell to threshold. Not only does the low input resistance make individual excitatory postsynaptic potentials brief so that they must be generated within 1 msec to sum but also the voltage-sensitive conductances of octopus cells prevent firing if the activation of auditory nerve inputs is not sufficiently synchronous and depolarization is not sufficiently rapid. In vivo in cats, octopus cells can fire rapidly and respond with exceptionally well-timed action potentials to periodic, broadband sounds such as clicks. Thus both the anatomical specializations and the biophysical specializations make octopus cells detectors of the coincident firing of their auditory nerve fiber inputs.


BT1 : Psychophysiology. 2004 May;41(3):341-9.

The music of speech: music training facilitates pitch processing in both music and language.

Schon D, Magne C, Besson M.

The main aim of the present experiment was to determine whether extensive musical training facilitates pitch contour processing not only in music but also in language. We used a parametric manipulation of final notes' or words' fundamental frequency (F0), and we recorded behavioral and electrophysiological data to examine the precise time course of pitch processing. We compared professional musicians and nonmusicians. Results revealed that within both domains, musicians detected weak F0 manipulations better than nonmusicians. Moreover, F0 manipulations within both music and language elicited similar variations in brain electrical potentials, with overall shorter onset latency for musicians than for nonmusicians. Finally, the scalp distribution of an early negativity in the linguistic task varied with musical expertise, being largest over temporal sites bilaterally for musicians and largest centrally and over left temporal sites for nonmusicians. These results are taken as evidence that extensive musical training influences the perception of pitch contour in spoken language.