Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 18 de 18
Filter
Add more filters










Publication year range
1.
J Assoc Res Otolaryngol ; 23(1): 75-94, 2022 02.
Article in English | MEDLINE | ID: mdl-34904205

ABSTRACT

Spatial hearing facilitates the perceptual organization of complex soundscapes into accurate mental representations of sound sources in the environment. Yet, the role of binaural cues in auditory scene analysis (ASA) has received relatively little attention in recent neuroscientific studies employing novel, spectro-temporally complex stimuli. This may be because a stimulation paradigm that provides binaurally derived grouping cues of sufficient spectro-temporal complexity has not yet been established for neuroscientific ASA experiments. Random-chord stereograms (RCS) are a class of auditory stimuli that exploit spectro-temporal variations in the interaural envelope correlation of noise-like sounds with interaurally coherent fine structure; they evoke salient auditory percepts that emerge only under binaural listening. Here, our aim was to assess the usability of the RCS paradigm for indexing binaural processing in the human brain. To this end, we recorded EEG responses to RCS stimuli from 12 normal-hearing subjects. The stimuli consisted of an initial 3-s noise segment with interaurally uncorrelated envelopes, followed by another 3-s segment, where envelope correlation was modulated periodically according to the RCS paradigm. Modulations were applied either across the entire stimulus bandwidth (wideband stimuli) or in temporally shifting frequency bands (ripple stimulus). Event-related potentials and inter-trial phase coherence analyses of the EEG responses showed that the introduction of the 3- or 5-Hz wideband modulations produced a prominent change-onset complex and ongoing synchronized responses to the RCS modulations. In contrast, the ripple stimulus elicited a change-onset response but no response to ongoing RCS modulation. Frequency-domain analyses revealed increased spectral power at the fundamental frequency and the first harmonic of wideband RCS modulations. RCS stimulation yields robust EEG measures of binaurally driven auditory reorganization and has potential to provide a flexible stimulation paradigm suitable for isolating binaural effects in ASA experiments.


Subject(s)
Auditory Perception , Cues , Acoustic Stimulation , Auditory Perception/physiology , Electroencephalography , Hearing/physiology , Humans
2.
Hear Res ; 377: 196-207, 2019 06.
Article in English | MEDLINE | ID: mdl-30981050

ABSTRACT

Making small head movements facilitates spatial hearing by resolving front-back confusions, otherwise common in free field sound source localization. The changes in interaural time difference (ITD) in response to head rotation provide a robust front-back cue, but whether interaural level difference (ILD) can be used as a dynamic cue is not clear. Therefore, the purpose of the present study was to assess the usefulness of dynamic ILD as a localization cue. The results show that human listeners were capable of correctly indicating the front-back dimension of high-frequency sinusoids based on level dynamics in free field conditions, but only if a wide movement range was allowed (±40∘). When the free field conditions were replaced by simplistic headphone stimulation, front-back responses were in agreement with the simulated source directions even with relatively small movement ranges (±5∘), whenever monaural sound level and ILD changed monotonically in response to head rotation. In conclusion, human listeners can use level dynamics as a front-back localization cue when the dynamics are monotonic. However, in free field conditions and particularly for narrowband target signals, this is often not the case. Therefore, the primary limiting factor in the use of dynamic level cues resides in the acoustic domain behavior of the cue itself, rather than in potential processing limitations or strategies of the human auditory system.


Subject(s)
Auditory Pathways/physiology , Cues , Head Movements , Sound Localization , Acoustic Stimulation , Adult , Female , Humans , Male , Motion , Sound , Time Factors , Young Adult
3.
Neuropsychologia ; 109: 181-199, 2018 01 31.
Article in English | MEDLINE | ID: mdl-29247667

ABSTRACT

Perceptual prominence of linguistic units such as words has been earlier connected to the concepts of predictability and attentional orientation. One hypothesis is that low-probability prosodic or lexical content is perceived as prominent due to the surprisal and high information value associated with the stimulus. However, the existing behavioral studies have used stimulus manipulations that follow or violate typical linguistic patterns present in the listeners' native language, i.e., assuming that the listeners have already established a model for acceptable prosodic patterns in the language. In the present study, we investigated whether prosodic expectations and the resulting subjective impression of prominence is affected by brief statistical adaptation to suprasegmental acoustic features in speech, also in the case where the prosodic patterns do not necessarily follow language-typical marking for prominence. We first exposed listeners to five minutes of speech with uneven distributions of falling and rising fundamental frequency (F0) trajectories on sentence-final words, and then tested their judgments of prominence on a set of new utterances. The results show that the probability of the F0 trajectory affects the perception of prominence, a less frequent F0 trajectory making a word more prominent independently of the absolute direction of F0 change. In the second part of the study, we conducted EEG-measurements on a set of new subjects listening to similar utterances with predominantly rising or falling F0 on sentence-final words. Analysis of the resulting event-related potentials (ERP) reveals a significant difference in N200 and N400 ERP-component amplitudes between standard and deviant prosody, again independently of the F0 direction and the underlying lexical content. Since N400 has earlier been associated with semantic processing of stimuli, this suggests that listeners implicitly track probabilities at the suprasegmental level and that predictability of a prosodic pattern during a word has an impact to the semantic processing of the word. Overall, the study suggests that prosodic markers for prominence are at least partially driven by the statistical structure of recently perceived speech, and therefore prominence perception could be based on statistical learning mechanisms similar to those observed in early word learning, but in this case operating at the level of suprasegmental acoustic features.


Subject(s)
Anticipation, Psychological/physiology , Brain/physiology , Psycholinguistics , Semantics , Speech Acoustics , Speech Perception/physiology , Adult , Electroencephalography , Evoked Potentials , Female , Humans , Judgment/physiology , Male , Models, Psychological , Probability Learning
4.
Neuroimage ; 167: 95-103, 2018 02 15.
Article in English | MEDLINE | ID: mdl-29122721

ABSTRACT

Binaural hearing, the ability to detect small differences in the timing and level of sounds at the two ears, underpins the ability to localize sound sources along the horizontal plane, and is important for decoding complex spatial listening environments into separate objects - a critical factor in 'cocktail-party listening'. For human listeners, the most important spatial cue is the interaural time difference (ITD). Despite many decades of neurophysiological investigations of ITD sensitivity in small mammals, and computational models aimed at accounting for human perception, a lack of concordance between these studies has hampered our understanding of how the human brain represents and processes ITDs. Further, neural coding of spatial cues might depend on factors such as head-size or hearing range, which differ considerably between humans and commonly used experimental animals. Here, using magnetoencephalography (MEG) in human listeners, and electro-corticography (ECoG) recordings in guinea pig-a small mammal representative of a range of animals in which ITD coding has been assessed at the level of single-neuron recordings-we tested whether processing of ITDs in human auditory cortex accords with a frequency-dependent periodic code of ITD reported in small mammals, or whether alternative or additional processing stages implemented in psychoacoustic models of human binaural hearing must be assumed. Our data were well accounted for by a model consisting of periodically tuned ITD-detectors, and were highly consistent across the two species. The results suggest that the representation of ITD in human auditory cortex is similar to that found in other mammalian species, a representation in which neural responses to ITD are determined by phase differences relative to sound frequency rather than, for instance, the range of ITDs permitted by head size or the absolute magnitude or direction of ITD.


Subject(s)
Auditory Cortex/physiology , Electrocorticography/methods , Magnetoencephalography/methods , Sound Localization/physiology , Adult , Animals , Female , Guinea Pigs , Humans , Male , Psychoacoustics , Time Factors , Young Adult
5.
Hear Res ; 327: 143-52, 2015 Sep.
Article in English | MEDLINE | ID: mdl-26074304

ABSTRACT

Human sound source localization relies on acoustical cues, most importantly, the interaural differences in time and level (ITD and ILD). For reaching a unified representation of auditory space the auditory nervous system needs to combine the information provided by these two cues. In search for such a unified representation, we conducted a magnetoencephalography (MEG) experiment that took advantage of the location-specific adaptation of the auditory cortical N1 response. In general, the attenuation caused by a preceding adaptor sound to the response elicited by a probe depends on their spatial arrangement: if the two sounds coincide, adaptation is stronger than when the locations differ. Here, we presented adaptor-probe pairs that contained different localization cues, for instance, adaptors with ITD and probes with ILD. We found that the adaptation of the N1 amplitude was location-specific across localization cues. This result can be explained by the existence of auditory cortical neurons that are sensitive to sound source location independent on which cue, ITD or ILD, provides the location information. Such neurons would form a cue-independent, unified representation of auditory space in human auditory cortex.


Subject(s)
Auditory Cortex/physiology , Cues , Sound Localization , Space Perception , Acoustic Stimulation , Adaptation, Psychological , Adult , Audiometry , Auditory Pathways/physiology , Auditory Threshold , Evoked Potentials, Auditory , Female , Humans , Magnetoencephalography , Male , Psychoacoustics , Time Factors
6.
J Acoust Soc Am ; 137(6): 3356-65, 2015 Jun.
Article in English | MEDLINE | ID: mdl-26093425

ABSTRACT

Natural auditory scenes often consist of several sound sources overlapping in time, but separated in space. Yet, location is not fully exploited in auditory grouping: spatially separated sounds can get perceptually fused into a single auditory object and this leads to difficulties in the identification and localization of concurrent sounds. Here, the brain mechanisms responsible for grouping across spatial locations were explored in magnetoencephalography (MEG) recordings. The results show that the cortical representation of a vowel spatially separated into two locations reflects the perceived location of the speech sound rather than the physical locations of the individual components. In other words, the auditory scene is neurally rearranged to bring components into spatial alignment when they were deemed to belong to the same object. This renders the original spatial information unavailable at the level of the auditory cortex and may contribute to difficulties in concurrent sound segregation.


Subject(s)
Auditory Cortex/physiology , Auditory Pathways/physiology , Sound Localization , Speech Acoustics , Speech Perception , Voice Quality , Acoustic Stimulation , Humans , Magnetoencephalography , Male , Psychoacoustics , Signal Detection, Psychological , Sound Spectrography
7.
Biol Psychol ; 109: 103-10, 2015 Jul.
Article in English | MEDLINE | ID: mdl-25960070

ABSTRACT

Functional brain imaging has identified specialized neural systems supporting human body perception. Responses to nude vs. clothed bodies within this system are amplified. However, it remains unresolved whether nude and clothed bodies are processed by same cerebral networks or whether processing of nude bodies recruits additional affective and arousal processing areas. We recorded simultaneous MEG and EEG while participants viewed photographs of clothed and nude bodies. Global field power revealed a peak ∼145ms after stimulus onset to both clothed and nude bodies, and ∼205ms exclusively to nude bodies. Nude-body-sensitive responses were centered first (100-200ms) in the extrastriate and fusiform body areas, and subsequently (200-300ms) in affective-motivational areas including insula and anterior cingulate cortex. We conclude that visibility of sexual features facilitates early cortical processing of human bodies, the purpose of which is presumably to trigger sexual behavior and ultimately ensure reproduction.


Subject(s)
Cerebral Cortex/physiology , Electroencephalography/methods , Human Body , Magnetoencephalography/methods , Sexuality/physiology , Visual Perception/physiology , Adult , Brain Mapping , Female , Humans , Male , Young Adult
8.
Hear Res ; 323: 99-106, 2015 May.
Article in English | MEDLINE | ID: mdl-25668126

ABSTRACT

Human sound source localization relies on various acoustical cues one of the most important being the interaural time difference (ITD). ITD is best detected in the fine structure of low-frequency sounds but it may also contribute to spatial hearing at higher frequencies if extracted from the sound envelope. The human brain mechanisms related to this envelope ITD cue remain unexplored. Here, we tested the sensitivity of the human auditory cortex to envelope ITD in magnetoencephalography (MEG) recordings. We found two types of sensitivity to envelope ITD. First, the amplitude of the auditory cortical N1m response was smaller for zero envelope ITD than for long envelope ITDs corresponding to the sound being in opposite phase in the two ears. Second, the N1m response amplitude showed ITD-specific adaptation for both fine-structure and for envelope ITD. The auditory cortical sensitivity was weaker for envelope ITD in high-frequency sounds than for fine-structure ITD in low-frequency sounds but occurred within a range of ITDs that are encountered in natural conditions. Finally, the participants were briefly tested for their behavioral ability to detect envelope ITD. Interestingly, we found a correlation between the behavioral performance and the neural sensitivity to envelope ITD. In conclusion, our findings show that the human auditory cortex is sensitive to ITD in the envelope of high-frequency sounds and this sensitivity may have behavioral relevance.


Subject(s)
Auditory Cortex/physiology , Cues , Sound Localization , Acoustic Stimulation , Adult , Audiometry, Pure-Tone , Female , Humans , Magnetoencephalography , Male , Psychoacoustics , Time Factors , Young Adult
9.
J Acoust Soc Am ; 137(2): EL190-3, 2015 Feb.
Article in English | MEDLINE | ID: mdl-25698049

ABSTRACT

Interaural level difference (ILD) is used as a cue in horizontal sound source localization. In free field, the magnitude of ILD depends on frequency: it is more prominent at high than low frequencies. Here, a magnetoencephalography experiment was conducted to test whether the sensitivity of the human auditory cortex to ILD is also frequency-dependent. Robust cortical sensitivity to ILD was found that could not be explained by monaural level effects, but this sensitivity did not differ between low- and high-frequency stimuli. This is consistent with previous psychoacoustical investigations showing that performance in ILD discrimination is not dependent on frequency.


Subject(s)
Acoustic Stimulation/methods , Auditory Cortex/physiology , Auditory Threshold , Cues , Sound Localization , Adult , Audiometry , Electrooculography , Eye Movements , Female , Humans , Magnetoencephalography , Male , Motion , Pressure , Signal Detection, Psychological , Sound , Time Factors , Young Adult
10.
Proc Natl Acad Sci U S A ; 110(37): 15151-6, 2013 Sep 10.
Article in English | MEDLINE | ID: mdl-23980161

ABSTRACT

The ability to locate the direction of a target sound in a background of competing sources is critical to the survival of many species and important for human communication. Nevertheless, brain mechanisms that provide for such accurate localization abilities remain poorly understood. In particular, it remains unclear how the auditory brain is able to extract reliable spatial information directly from the source when competing sounds and reflections dominate all but the earliest moments of the sound wave reaching each ear. We developed a stimulus mimicking the mutual relationship of sound amplitude and binaural cues, characteristic to reverberant speech. This stimulus, named amplitude modulated binaural beat, allows for a parametric and isolated change of modulation frequency and phase relations. Employing magnetoencephalography and psychoacoustics it is demonstrated that the auditory brain uses binaural information in the stimulus fine structure only during the rising portion of each modulation cycle, rendering spatial information recoverable in an otherwise unlocalizable sound. The data suggest that amplitude modulation provides a means of "glimpsing" low-frequency spatial cues in a manner that benefits listening in noisy or reverberant environments.


Subject(s)
Sound Localization/physiology , Acoustic Stimulation , Adult , Auditory Cortex/physiology , Female , Hearing/physiology , Humans , Magnetoencephalography , Male , Psychoacoustics , Time Factors , Young Adult
11.
Front Neurosci ; 7: 44, 2013.
Article in English | MEDLINE | ID: mdl-23543781

ABSTRACT

The auditory cortex represents spatial locations differently from other sensory modalities. While visual and tactile cortices utilize topographical space maps, for audition no such cortical map has been found. Instead, auditory cortical neurons have wide spatial receptive fields and together they form a population rate code of sound source location. Recent studies have shown that this code is modulated by task conditions so that during auditory tasks it provides better selectivity to sound source location than during idle listening. The goal of this study was to establish whether the neural representation of auditory space can also be influenced by task conditions involving other sensory modalities than hearing. Therefore, we conducted magnetoencephalography (MEG) recordings in which auditory spatial selectivity of the human cortex was probed with an adaptation paradigm while subjects performed a visual task. Engaging in the task led to an increase in neural selectivity to sound source location compared to when no task was performed. This suggests that an enhancement in the population rate code of auditory space took place during task performance. This enhancement in auditory spatial selectivity was independent of the direction of visual orientation. Together with previous studies, these findings suggest that performing any demanding task, even one in which sounds and their source locations are irrelevant, can lead to enhancements in the neural representation of auditory space. Such mechanisms may have great survival value as sounds are capable of producing location information on potentially relevant events in all directions and over long distances.

12.
Neuroscientist ; 18(6): 602-12, 2012 Dec.
Article in English | MEDLINE | ID: mdl-22492193

ABSTRACT

The auditory system codes spatial locations in a way that deviates from the spatial representations found in other modalities. This difference is especially striking in the cortex, where neurons form topographical maps of visual and tactile space but where auditory space is represented through a population rate code. In this hemifield code, sound source location is represented in the activity of two widely tuned opponent populations, one tuned to the right and the other to the left side of auditory space. Scientists are only beginning to uncover how this coding strategy adapts to various spatial processing demands. This review presents the current understanding of auditory spatial processing in the cortex. To this end, the authors consider how various implementations of the hemifield code may exist within the auditory cortex and how these may be modulated by the stimulation and task context. As a result, a coherent set of neural strategies for auditory spatial processing emerges.


Subject(s)
Auditory Cortex/physiology , Sound Localization/physiology , Space Perception/physiology , Auditory Perception/physiology , Functional Laterality/physiology , Humans
13.
Brain Res ; 1367: 298-309, 2011 Jan 07.
Article in English | MEDLINE | ID: mdl-20969833

ABSTRACT

The cortical mechanisms underlying human speech perception in acoustically adverse conditions remain largely unknown. Besides distortions from external sources, degradation of the acoustic structure of the sound itself poses further demands on perceptual mechanisms. We conducted a magnetoencephalography (MEG) study to reveal whether the perceptual differences between these distortions are reflected in cortically generated auditory evoked fields (AEFs). To mimic the degradation of the internal structure of sound and external distortion, we degraded speech sounds by reducing the amplitude resolution of the signal waveform and by using additive noise, respectively. Since both distortion types increase the relative strength of high frequencies in the signal spectrum, we also used versions of the stimuli which were low-pass filtered to match the tilted spectral envelope of the undistorted speech sound. This enabled us to examine whether the changes in the overall spectral shape of the stimuli affect the AEFs. We found that the auditory N1m response was substantially enhanced as the amplitude resolution was reduced. In contrast, the N1m was insensitive to distorted speech with additive noise. Changing the spectral envelope had no effect on the N1m. We propose that the observed amplitude enhancements are due to an increase in noisy spectral harmonics produced by the reduction of the amplitude resolution, which activates the periodicity-sensitive neuronal populations participating in pitch extraction processes. The current findings suggest that the auditory cortex processes speech sounds in a differential manner when the internal structure of sound is degraded compared with the speech distorted by external noise.


Subject(s)
Auditory Cortex/physiology , Brain Waves/physiology , Evoked Potentials, Auditory/physiology , Noise , Sound Localization/physiology , Speech Perception/physiology , Acoustic Stimulation , Adult , Brain Mapping , Humans , Magnetoencephalography , Male , Phonetics , Psychoacoustics , Reaction Time/physiology , Spectrum Analysis , Statistics, Nonparametric , Young Adult
14.
J Acoust Soc Am ; 127(2): EL60-5, 2010 Feb.
Article in English | MEDLINE | ID: mdl-20136180

ABSTRACT

A magnetoencephalography study was conducted to reveal the neural code of interaural time difference (ITD) in the human cortex. Widely used crosscorrelator models predict that the code consists of narrow receptive fields distributed to all ITDs. The present findings are, however, more in line with a neural code formed by two opponent neural populations: one tuned to the left and the other to the right hemifield. The results are consistent with models of ITD extraction in the auditory brainstem of small mammals and, therefore, suggest that similar computational principles underlie human sound source localization.


Subject(s)
Auditory Cortex/physiology , Auditory Perception/physiology , Ear , Models, Neurological , Acoustic Stimulation , Adult , Evoked Potentials, Auditory , Female , Functional Laterality , Head , Humans , Magnetoencephalography , Male , Photic Stimulation , Time Factors , Visual Perception/physiology
15.
Brain Res ; 1306: 93-9, 2010 Jan 08.
Article in English | MEDLINE | ID: mdl-19799877

ABSTRACT

Recent single-neuron recordings in monkeys and magnetoencephalography (MEG) data on humans suggest that auditory space is represented in cortex as a population rate code whereby spatial receptive fields are wide and centered at locations to the far left or right of the subject. To explore the details of this code in the human brain, we conducted an MEG study utilizing realistic spatial sound stimuli presented in a stimulus-specific adaptation paradigm. In this paradigm, the spatial selectivity of cortical neurons is measured as the effect the location of a preceding adaptor has on the response to a subsequent probe sound. Two types of stimuli were used: a wideband noise sound and a speech sound. The cortical hemispheres differed in the effects the adaptors had on the response to a probe sound presented in front of the subject. The right-hemispheric responses were attenuated more by an adaptor to the left than by an adaptor to the right of the subject. In contrast, the left-hemispheric responses were similarly affected by adaptors in these two locations. When interpreted in terms of single-neuron spatial receptive fields, these results support a population rate code model where neurons in the right hemisphere are more often tuned to the left than to the right of the perceiver while in the left hemisphere these two neuronal populations are of equal size.


Subject(s)
Auditory Perception/physiology , Cerebral Cortex/physiology , Sound Localization/physiology , Space Perception/physiology , Acoustic Stimulation , Adult , Analysis of Variance , Evoked Potentials, Auditory , Female , Functional Laterality , Humans , Magnetoencephalography , Male , Models, Neurological , Neurons/physiology , Speech , Speech Perception/physiology
16.
PLoS One ; 4(10): e7600, 2009 Oct 26.
Article in English | MEDLINE | ID: mdl-19855836

ABSTRACT

BACKGROUND: Previous work on the human auditory cortex has revealed areas specialized in spatial processing but how the neurons in these areas represent the location of a sound source remains unknown. METHODOLOGY/PRINCIPAL FINDINGS: Here, we performed a magnetoencephalography (MEG) experiment with the aim of revealing the neural code of auditory space implemented by the human cortex. In a stimulus-specific adaptation paradigm, realistic spatial sound stimuli were presented in pairs of adaptor and probe locations. We found that the attenuation of the N1m response depended strongly on the spatial arrangement of the two sound sources. These location-specific effects showed that sounds originating from locations within the same hemifield activated the same neuronal population regardless of the spatial separation between the sound sources. In contrast, sounds originating from opposite hemifields activated separate groups of neurons. CONCLUSIONS/SIGNIFICANCE: These results are highly consistent with a rate code of spatial location formed by two opponent populations, one tuned to locations in the left and the other to those in the right. This indicates that the neuronal code of sound source location implemented by the human auditory cortex is similar to that previously found in other primates.


Subject(s)
Auditory Cortex/anatomy & histology , Magnetoencephalography/methods , Sound Localization/physiology , Acoustic Stimulation/methods , Adult , Auditory Cortex/physiology , Auditory Pathways/physiology , Auditory Perception/physiology , Brain Mapping , Data Interpretation, Statistical , Evoked Potentials, Auditory/physiology , Humans , Neurons/metabolism , Sound
17.
Cogn Affect Behav Neurosci ; 9(3): 304-13, 2009 Sep.
Article in English | MEDLINE | ID: mdl-19679765

ABSTRACT

Our native language has a lifelong effect on how we perceive speech sounds. Behaviorally, this is manifested as categorical perception, but the neural mechanisms underlying this phenomenon are still unknown. Here, we constructed a computational model of categorical perception, following principles consistent with infant speech learning. A self-organizing network was exposed to a statistical distribution of speech input presented as neural activity patterns of the auditory periphery, resembling the way sound arrives to the human brain. In the resulting neural map, categorical perception emerges from most single neurons of the model being maximally activated by prototypical speech sounds, while the largest variability in activity is produced at category boundaries. Consequently, regions in the vicinity of prototypes become perceptually compressed, and regions at category boundaries become expanded. Thus, the present study offers a unifying framework for explaining the neural basis of the warping of perceptual space associated with categorical perception.


Subject(s)
Brain/physiology , Mental Processes/physiology , Models, Neurological , Neural Networks, Computer , Speech Perception/physiology , Child Language , Humans , Infant , Learning/physiology , Neurons/physiology , Phonetics , Sound Spectrography , Speech
18.
Neurosci Lett ; 396(1): 17-22, 2006 Mar 20.
Article in English | MEDLINE | ID: mdl-16343772

ABSTRACT

In an attempt to delineate the assumed 'what' and 'where' processing streams, we studied the processing of spatial sound in the human cortex by using magnetoencephalography in the passive and active recording conditions and two kinds of spatial stimuli: individually constructed, highly realistic spatial (3D) stimuli and stimuli containing interaural time difference (ITD) cues only. The auditory P1m, N1m, and P2m responses of the event-related field were found to be sensitive to the direction of sound source in the azimuthal plane. In general, the right-hemispheric responses to spatial sounds were more prominent than the left-hemispheric ones. The right-hemispheric P1m and N1m responses peaked earlier for sound sources in the contralateral than for sources in the ipsilateral hemifield and the peak amplitudes of all responses reached their maxima for contralateral sound sources. The amplitude of the right-hemispheric P2m response reflected the degree of spatiality of sound, being twice as large for the 3D than ITD stimuli. The results indicate that the right hemisphere is specialized in the processing of spatial cues in the passive recording condition. Minimum current estimate (MCE) localization revealed that temporal areas were activated both in the active and passive condition. This initial activation, taking place at around 100 ms, was followed by parietal and frontal activity at 180 and 200 ms, respectively. The latter activations, however, were specific to attentional engagement and motor responding. This suggests that parietal activation reflects active responding to a spatial sound rather than auditory spatial processing as such.


Subject(s)
Auditory Cortex/physiology , Evoked Potentials, Auditory/physiology , Pitch Discrimination/physiology , Reaction Time/physiology , Sound Localization/physiology , Space Perception/physiology , Acoustic Stimulation , Adult , Auditory Cortex/anatomy & histology , Auditory Pathways/anatomy & histology , Auditory Pathways/physiology , Frontal Lobe/anatomy & histology , Frontal Lobe/physiology , Functional Laterality/physiology , Humans , Magnetoencephalography , Male , Nerve Net/anatomy & histology , Nerve Net/physiology , Neural Pathways/anatomy & histology , Neural Pathways/physiology , Parietal Lobe/anatomy & histology , Parietal Lobe/physiology
SELECTION OF CITATIONS
SEARCH DETAIL
...