Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 103
Filter
Add more filters










Publication year range
1.
bioRxiv ; 2024 Mar 13.
Article in English | MEDLINE | ID: mdl-38586037

ABSTRACT

Hearing-impaired listeners struggle to understand speech in noise, even when using cochlear implants (CIs) or hearing aids. Successful listening in noisy environments depends on the brain's ability to organize a mixture of sound sources into distinct perceptual streams (i.e., source segregation). In normal-hearing listeners, temporal coherence of sound fluctuations across frequency channels supports this process by promoting grouping of elements belonging to a single acoustic source. We hypothesized that reduced spectral resolution-a hallmark of both electric/CI (from current spread) and acoustic (from broadened tuning) hearing with sensorineural hearing loss-degrades segregation based on temporal coherence. This is because reduced frequency resolution decreases the likelihood that a single sound source dominates the activity driving any specific channel; concomitantly, it increases the correlation in activity across channels. Consistent with our hypothesis, predictions from a physiologically plausible model of temporal-coherence-based segregation suggest that CI current spread reduces comodulation masking release (CMR; a correlate of temporal-coherence processing) and speech intelligibility in noise. These predictions are consistent with our behavioral data with simulated CI listening. Our model also predicts smaller CMR with increasing levels of outer-hair-cell damage. These results suggest that reduced spectral resolution relative to normal hearing impairs temporal-coherence-based segregation and speech-in-noise outcomes.

2.
JASA Express Lett ; 4(3)2024 03 01.
Article in English | MEDLINE | ID: mdl-38526127

ABSTRACT

Listeners performed two different tasks in which they remembered short sequences comprising either complex tones (generally heard as one melody) or everyday sounds (generally heard as separate objects). In one, listeners judged whether a probe item had been present in the preceding sequence. In the other, they judged whether a second sequence of the same items was identical in order to the preceding sequence. Performance on the first task was higher for everyday sounds; performance on the second was higher for complex tones. Perceptual organization strongly shapes listeners' memory for sounds, with implications for real-world communication.


Subject(s)
Auditory Perception , Memory, Short-Term , Sound , Hearing , Communication
3.
Autism Res ; 16(10): 1859-1876, 2023 10.
Article in English | MEDLINE | ID: mdl-37735966

ABSTRACT

Limited research has evaluated neural encoding of sounds from a developmental perspective in individuals with autism (ASD), especially among those with intellectual disability. We compared auditory evoked potentials (AEPs) in autistic adolescents with a wide range of intellectual abilities (n = 40, NVIQ 30-160) to both age-matched cognitively able neurotypical adolescent controls (NT-A, n = 37) and younger neurotypical children (NT-C, n = 27) to assess potential developmental delays. In addition to a classic measure of peak amplitude, we calculated a continuous measure of intra-class correlation (ICC) between each adolescent participant's AEP and the age-normative, average AEP waveforms calculated from NT-C and NT-A to study differences in signal morphology. We found that peak amplitudes of neural responses were significantly smaller in autistic adolescents compared to NT-A. We also found that the AEP morphology of autistic adolescents looked more like NT-A peers than NT-C but was still significantly different from NT-A AEP waveforms. Results suggest that AEPs of autistic adolescents present differently from NTs, regardless of age, and differences cannot be accounted for by developmental delay. Nonverbal intelligence significantly predicted how closely each adolescent's AEP resembled the age-normed waveform. These results support an evolving theory that the degree of disruption in early neural responses to low-level inputs is reflected in the severity of intellectual impairments in autism.


Subject(s)
Autism Spectrum Disorder , Autistic Disorder , Child , Humans , Adolescent , Autism Spectrum Disorder/complications , Evoked Potentials, Auditory/physiology , Sound , Brain/physiology , Evoked Potentials
4.
J Biomed Opt ; 28(7): 075001, 2023 07.
Article in English | MEDLINE | ID: mdl-37457628

ABSTRACT

Significance: Using functional near-infrared spectroscopy (fNIRS) in bottlenose dolphins (Tursiops truncatus) could help to understand how echolocating animals perceive their environment and how they focus on specific auditory objects, such as fish, in noisy marine settings. Aim: To test the feasibility of near-infrared spectroscopy (NIRS) in medium-sized marine mammals, such as dolphins, we modeled the light propagation with computational tools to determine the wavelengths, optode locations, and separation distances that maximize sensitivity to brain tissue. Approach: Using frequency-domain NIRS, we measured the absorption and reduced scattering coefficient of dolphin sculp. We assigned muscle, bone, and brain optical properties from the literature and modeled light propagation in a spatially accurate and biologically relevant model of a dolphin head, using finite-element modeling. We assessed tissue sensitivities for a range of wavelengths (600 to 1700 nm), source-detector distances (50 to 120 mm), and animal sizes (juvenile model 25% smaller than adult). Results: We found that the wavelengths most suitable for imaging the brain fell into two ranges: 700 to 900 nm and 1100 to 1150 nm. The optimal location for brain sensing positioned the center point between source and detector 30 to 50 mm caudal of the blowhole and at an angle 45 deg to 90 deg lateral off the midsagittal plane. Brain tissue sensitivity comparable to human measurements appears achievable only for smaller animals, such as juvenile bottlenose dolphins or smaller species of cetaceans, such as porpoises, or with source-detector separations ≫100 mm in adult dolphins. Conclusions: Brain measurements in juvenile or subadult dolphins, or smaller dolphin species, may be possible using specialized fNIRS devices that support optode separations of >100 mm. We speculate that many measurement repetitions will be required to overcome hemodynamic signals originating predominantly from the muscle layer above the skull. NIRS measurements of muscle tissue are feasible today with source-detector separations of 50 mm, or even less.


Subject(s)
Bottle-Nosed Dolphin , Humans , Animals , Adult , Bottle-Nosed Dolphin/physiology , Spectroscopy, Near-Infrared , Feasibility Studies , Head
5.
Sci Rep ; 13(1): 10216, 2023 06 23.
Article in English | MEDLINE | ID: mdl-37353552

ABSTRACT

Neurophysiological studies suggest that intrinsic brain oscillations influence sensory processing, especially of rhythmic stimuli like speech. Prior work suggests that brain rhythms may mediate perceptual grouping and selective attention to speech amidst competing sound, as well as more linguistic aspects of speech processing like predictive coding. However, we know of no prior studies that have directly tested, at the single-trial level, whether brain oscillations relate to speech-in-noise outcomes. Here, we combined electroencephalography while simultaneously measuring intelligibility of spoken sentences amidst two different interfering sounds: multi-talker babble or speech-shaped noise. We find that induced parieto-occipital alpha (7-15 Hz; thought to modulate attentional focus) and frontal beta (13-30 Hz; associated with maintenance of the current sensorimotor state and predictive coding) oscillations covary with trial-wise percent-correct scores; importantly, alpha and beta power provide significant independent contributions to predicting single-trial behavioral outcomes. These results can inform models of speech processing and guide noninvasive measures to index different neural processes that together support complex listening.


Subject(s)
Speech Intelligibility , Speech Perception , Speech Perception/physiology , Noise , Auditory Perception , Electroencephalography
6.
Cognition ; 238: 105473, 2023 09.
Article in English | MEDLINE | ID: mdl-37210878

ABSTRACT

Statistical learning across passive exposure has been theoretically situated with unsupervised learning. However, when input statistics accumulate over established representations - like speech syllables, for example - there is the possibility that prediction derived from activation of rich, existing representations may support error-driven learning. Here, across five experiments, we present evidence for error-driven learning across passive speech listening. Young adults passively listened to a string of eight beer - pier speech tokens with distributional regularities following either a canonical American-English acoustic dimension correlation or a correlation reversed to create an accent. A sequence-final test stimulus assayed the perceptual weight - the effectiveness - of the secondary dimension in signaling category membership as a function of preceding sequence regularities. Perceptual weight flexibly adjusted according to the passively experienced regularities even when the preceding regularities shifted on a trial-by-trial basis. The findings align with a theoretical view that activation of established internal representations can support learning across statistical regularities via error-driven learning. At the broadest level, this suggests that not all statistical learning need be unsupervised. Moreover, these findings help to account for how cognitive systems may accommodate competing demands for flexibility and stability: instead of overwriting existing representations when short-term input distributions depart from the norms, the mapping from input to category representations may be dynamically - and rapidly - adjusted via error-driven learning from predictions derived from internal representations.


Subject(s)
Speech Perception , Speech , Young Adult , Humans , Speech/physiology , Speech Perception/physiology , Auditory Perception , Language
7.
bioRxiv ; 2023 May 22.
Article in English | MEDLINE | ID: mdl-36712081

ABSTRACT

Neurophysiological studies suggest that intrinsic brain oscillations influence sensory processing, especially of rhythmic stimuli like speech. Prior work suggests that brain rhythms may mediate perceptual grouping and selective attention to speech amidst competing sound, as well as more linguistic aspects of speech processing like predictive coding. However, we know of no prior studies that have directly tested, at the single-trial level, whether brain oscillations relate to speech-in-noise outcomes. Here, we combined electroencephalography while simultaneously measuring intelligibility of spoken sentences amidst two different interfering sounds: multi-talker babble or speech-shaped noise. We find that induced parieto-occipital alpha (7-15 Hz; thought to modulate attentional focus) and frontal beta (13-30 Hz; associated with maintenance of the current sensorimotor state and predictive coding) oscillations covary with trial-wise percent-correct scores; importantly, alpha and beta power provide significant independent contributions to predicting single-trial behavioral outcomes. These results can inform models of speech processing and guide noninvasive measures to index different neural processes that together support complex listening.

8.
Wiley Interdiscip Rev Cogn Sci ; 14(1): e1610, 2023 Jan.
Article in English | MEDLINE | ID: mdl-35642475

ABSTRACT

Attention prioritizes certain information at the expense of other information in ways that are similar across vision, audition, and other sensory modalities. It influences how-and even what-information is represented and processed, affecting brain activity at every level. Much of the core research into cognitive and neural mechanisms of attention has used visual tasks. However, the same top-down, object-based, and bottom-up attentional processes shape auditory perception, largely through the same underlying, cognitive networks. This article is categorized under: Psychology > Attention.


Subject(s)
Auditory Perception , Magnetic Resonance Imaging , Humans , Visual Perception , Photic Stimulation
9.
Brain Res ; 1798: 148144, 2023 01 01.
Article in English | MEDLINE | ID: mdl-36328068

ABSTRACT

Human cognitive abilities naturally vary along a spectrum, even among those we call "neurotypical". Individuals differ in their ability to selectively attend to goal-relevant auditory stimuli. We sought to characterize this variability in a cohort of people with diverse attentional functioning. We recruited both neurotypical (N = 20) and ADHD (N = 25) young adults, all with normal hearing. Participants listened to one of three concurrent, spatially separated speech streams and reported the order of the syllables in that stream while we recorded electroencephalography (EEG). We tested both the ability to sustain attentional focus on a single "Target" stream and the ability to monitor the Target but flexibly either ignore or switch attention to an unpredictable "Interrupter" stream from another direction that sometimes appeared. Although differences in both stimulus structure and task demands affected behavioral performance, ADHD status did not. In both groups, the Interrupter evoked larger neural responses when it was to be attended compared to when it was irrelevant, including for the P3a "reorienting" response previously described as involuntary. This attentional modulation was weaker in ADHD listeners, even though their behavioral performance was the same. Across the entire cohort, individual performance correlated with the degree of top-down modulation of neural responses. These results demonstrate that listeners differ in their ability to modulate neural representations of sound based on task goals, while suggesting that adults with ADHD may have weaker volitional control of attentional processes than their neurotypical counterparts.


Subject(s)
Attention Deficit Disorder with Hyperactivity , Humans , Young Adult , Auditory Perception/physiology , Electroencephalography , Speech , Hearing Tests , Acoustic Stimulation
10.
Annu Int Conf IEEE Eng Med Biol Soc ; 2022: 760-763, 2022 07.
Article in English | MEDLINE | ID: mdl-36085807

ABSTRACT

Transcranial alternating current stimulation (tACS) is a neuromodulatory technique that is widely used to investigate the functions of oscillations in the brain. Despite increasing usage in both research and clinical settings, the mechanisms of tACS are still not completely understood. To shed light on these mechanisms, we injected alternating current into a Jansen and Rit neural mass model. Two cortical columns were linked with long-range connections to examine how alternating current impacted cortical connectivity. Alternating current injected to both columns increased power and coherence at the stimulation frequency; however this effect was greatest at the model's resonant frequency. Varying the phase of stimulation impacted the time it took for entrainment to stabilize, an effect we believe is due to constructive and destructive inteference with endogenous membrane currents. The power output the model also depended on the phase of the stimulation between cortical columns. These results provide insight on the mechanisms of neurostimulation, by demonstrating that tACS increases both power and coherence at a neural network's resonant frequency, in a phase-dependent manner.


Subject(s)
Electricity , Transcranial Direct Current Stimulation , Brain
11.
J Acoust Soc Am ; 151(5): 3219, 2022 05.
Article in English | MEDLINE | ID: mdl-35649920

ABSTRACT

Salient interruptions draw attention involuntarily. Here, we explored whether this effect depends on the spatial and temporal relationships between a target stream and interrupter. In a series of online experiments, listeners focused spatial attention on a target stream of spoken syllables in the presence of an otherwise identical distractor stream from the opposite hemifield. On some random trials, an interrupter (a cat "MEOW") occurred. Experiment 1 established that the interrupter, which occurred randomly in 25% of the trials in the hemifield opposite the target, degraded target recall. Moreover, a majority of participants exhibited this degradation for the first target syllable, which finished before the interrupter began. Experiment 2 showed that the effect of an interrupter was similar whether it occurred in the opposite or the same hemifield as the target. Experiment 3 found that the interrupter degraded performance slightly if it occurred before the target stream began but had no effect if it began after the target stream ended. Experiment 4 showed decreased interruption effects when the interruption frequency increased (50% of the trials). These results demonstrate that a salient interrupter disrupts recall of a target stream, regardless of its direction, especially if it occurs during a target stream.


Subject(s)
Mental Recall , Humans
12.
Cereb Cortex ; 32(4): 855-869, 2022 02 08.
Article in English | MEDLINE | ID: mdl-34467399

ABSTRACT

Working memory (WM) supports the persistent representation of transient sensory information. Visual and auditory stimuli place different demands on WM and recruit different brain networks. Separate auditory- and visual-biased WM networks extend into the frontal lobes, but several challenges confront attempts to parcellate human frontal cortex, including fine-grained organization and between-subject variability. Here, we use differential intrinsic functional connectivity from 2 visual-biased and 2 auditory-biased frontal structures to identify additional candidate sensory-biased regions in frontal cortex. We then examine direct contrasts of task functional magnetic resonance imaging during visual versus auditory 2-back WM to validate those candidate regions. Three visual-biased and 5 auditory-biased regions are robustly activated bilaterally in the frontal lobes of individual subjects (N = 14, 7 women). These regions exhibit a sensory preference during passive exposure to task stimuli, and that preference is stronger during WM. Hierarchical clustering analysis of intrinsic connectivity among novel and previously identified bilateral sensory-biased regions confirms that they functionally segregate into visual and auditory networks, even though the networks are anatomically interdigitated. We also observe that the frontotemporal auditory WM network is highly selective and exhibits strong functional connectivity to structures serving non-WM functions, while the frontoparietal visual WM network hierarchically merges into the multiple-demand cognitive system.


Subject(s)
Auditory Perception , Memory, Short-Term , Brain Mapping/methods , Female , Frontal Lobe/diagnostic imaging , Humans , Magnetic Resonance Imaging
13.
Ear Hear ; 43(1): 9-22, 2022.
Article in English | MEDLINE | ID: mdl-34751676

ABSTRACT

Following a conversation in a crowded restaurant or at a lively party poses immense perceptual challenges for some individuals with normal hearing thresholds. A number of studies have investigated whether noise-induced cochlear synaptopathy (CS; damage to the synapses between cochlear hair cells and the auditory nerve following noise exposure that does not permanently elevate hearing thresholds) contributes to this difficulty. A few studies have observed correlations between proxies of noise-induced CS and speech perception in difficult listening conditions, but many have found no evidence of a relationship. To understand these mixed results, we reviewed previous studies that have examined noise-induced CS and performance on speech perception tasks in adverse listening conditions in adults with normal or near-normal hearing thresholds. Our review suggests that superficially similar speech perception paradigms used in previous investigations actually placed very different demands on sensory, perceptual, and cognitive processing. Speech perception tests that use low signal-to-noise ratios and maximize the importance of fine sensory details- specifically by using test stimuli for which lexical, syntactic, and semantic cues do not contribute to performance-are more likely to show a relationship to estimated CS levels. Thus, the current controversy as to whether or not noise-induced CS contributes to individual differences in speech perception under challenging listening conditions may be due in part to the fact that many of the speech perception tasks used in past studies are relatively insensitive to CS-induced deficits.


Subject(s)
Speech Perception , Speech , Acoustic Stimulation , Adult , Auditory Threshold/physiology , Humans , Individuality , Perceptual Masking , Speech Perception/physiology
14.
J Neurosci ; 42(2): 240-254, 2022 01 12.
Article in English | MEDLINE | ID: mdl-34764159

ABSTRACT

Temporal coherence of sound fluctuations across spectral channels is thought to aid auditory grouping and scene segregation. Although prior studies on the neural bases of temporal-coherence processing focused mostly on cortical contributions, neurophysiological evidence suggests that temporal-coherence-based scene analysis may start as early as the cochlear nucleus (i.e., the first auditory region supporting cross-channel processing over a wide frequency range). Accordingly, we hypothesized that aspects of temporal-coherence processing that could be realized in early auditory areas may shape speech understanding in noise. We then explored whether physiologically plausible computational models could account for results from a behavioral experiment that measured consonant categorization in different masking conditions. We tested whether within-channel masking of target-speech modulations predicted consonant confusions across the different conditions and whether predictions were improved by adding across-channel temporal-coherence processing mirroring the computations known to exist in the cochlear nucleus. Consonant confusions provide a rich characterization of error patterns in speech categorization, and are thus crucial for rigorously testing models of speech perception; however, to the best of our knowledge, they have not been used in prior studies of scene analysis. We find that within-channel modulation masking can reasonably account for category confusions, but that it fails when temporal fine structure cues are unavailable. However, the addition of across-channel temporal-coherence processing significantly improves confusion predictions across all tested conditions. Our results suggest that temporal-coherence processing strongly shapes speech understanding in noise and that physiological computations that exist early along the auditory pathway may contribute to this process.SIGNIFICANCE STATEMENT Temporal coherence of sound fluctuations across distinct frequency channels is thought to be important for auditory scene analysis. Prior studies on the neural bases of temporal-coherence processing focused mostly on cortical contributions, and it was unknown whether speech understanding in noise may be shaped by across-channel processing that exists in earlier auditory areas. Using physiologically plausible computational modeling to predict consonant confusions across different listening conditions, we find that across-channel temporal coherence contributes significantly to scene analysis and speech perception and that such processing may arise in the auditory pathway as early as the brainstem. By virtue of providing a richer characterization of error patterns not obtainable with just intelligibility scores, consonant confusions yield unique insight into scene analysis mechanisms.


Subject(s)
Auditory Pathways/physiology , Auditory Perception/physiology , Cochlea/physiology , Speech/physiology , Acoustic Stimulation , Auditory Threshold/physiology , Humans , Models, Neurological , Perceptual Masking
15.
J Acoust Soc Am ; 150(4): 3085, 2021 10.
Article in English | MEDLINE | ID: mdl-34717460

ABSTRACT

The ability to see a talker's face improves speech intelligibility in noise, provided that the auditory and visual speech signals are approximately aligned in time. However, the importance of spatial alignment between corresponding faces and voices remains unresolved, particularly in multi-talker environments. In a series of online experiments, we investigated this using a task that required participants to selectively attend a target talker in noise while ignoring a distractor talker. In experiment 1, we found improved task performance when the talkers' faces were visible, but only when corresponding faces and voices were presented in the same hemifield (spatially aligned). In experiment 2, we tested for possible influences of eye position on this result. In auditory-only conditions, directing gaze toward the distractor voice reduced performance, but this effect could not fully explain the cost of audio-visual (AV) spatial misalignment. Lowering the signal-to-noise ratio (SNR) of the speech from +4 to -4 dB increased the magnitude of the AV spatial alignment effect (experiment 3), but accurate closed-set lipreading caused a floor effect that influenced results at lower SNRs (experiment 4). Taken together, these results demonstrate that spatial alignment between faces and voices contributes to the ability to selectively attend AV speech.


Subject(s)
Speech Perception , Voice , Humans , Lipreading , Noise/adverse effects , Speech Intelligibility
16.
J Acoust Soc Am ; 150(4): 2664, 2021 10.
Article in English | MEDLINE | ID: mdl-34717498

ABSTRACT

To understand the mechanisms of speech perception in everyday listening environments, it is important to elucidate the relative contributions of different acoustic cues in transmitting phonetic content. Previous studies suggest that the envelope of speech in different frequency bands conveys most speech content, while the temporal fine structure (TFS) can aid in segregating target speech from background noise. However, the role of TFS in conveying phonetic content beyond what envelopes convey for intact speech in complex acoustic scenes is poorly understood. The present study addressed this question using online psychophysical experiments to measure the identification of consonants in multi-talker babble for intelligibility-matched intact and 64-channel envelope-vocoded stimuli. Consonant confusion patterns revealed that listeners had a greater tendency in the vocoded (versus intact) condition to be biased toward reporting that they heard an unvoiced consonant, despite envelope and place cues being largely preserved. This result was replicated when babble instances were varied across independent experiments, suggesting that TFS conveys voicing information beyond what is conveyed by envelopes for intact speech in babble. Given that multi-talker babble is a masker that is ubiquitous in everyday environments, this finding has implications for the design of assistive listening devices such as cochlear implants.


Subject(s)
Cochlear Implants , Speech Perception , Acoustic Stimulation , Noise/adverse effects , Perceptual Masking , Phonetics , Speech , Speech Intelligibility
17.
J Acoust Soc Am ; 150(3): 2230, 2021 09.
Article in English | MEDLINE | ID: mdl-34598642

ABSTRACT

A fundamental question in the neuroscience of everyday communication is how scene acoustics shape the neural processing of attended speech sounds and in turn impact speech intelligibility. While it is well known that the temporal envelopes in target speech are important for intelligibility, how the neural encoding of target-speech envelopes is influenced by background sounds or other acoustic features of the scene is unknown. Here, we combine human electroencephalography with simultaneous intelligibility measurements to address this key gap. We find that the neural envelope-domain signal-to-noise ratio in target-speech encoding, which is shaped by masker modulations, predicts intelligibility over a range of strategically chosen realistic listening conditions unseen by the predictive model. This provides neurophysiological evidence for modulation masking. Moreover, using high-resolution vocoding to carefully control peripheral envelopes, we show that target-envelope coding fidelity in the brain depends not only on envelopes conveyed by the cochlea, but also on the temporal fine structure (TFS), which supports scene segregation. Our results are consistent with the notion that temporal coherence of sound elements across envelopes and/or TFS influences scene analysis and attentive selection of a target sound. Our findings also inform speech-intelligibility models and technologies attempting to improve real-world speech communication.


Subject(s)
Speech Intelligibility , Speech Perception , Acoustic Stimulation , Acoustics , Auditory Perception , Humans , Perceptual Masking , Signal-To-Noise Ratio
18.
Brain Lang ; 221: 104996, 2021 10.
Article in English | MEDLINE | ID: mdl-34358924

ABSTRACT

Speech is processed less efficiently from discontinuous, mixed talkers than one consistent talker, but little is known about the neural mechanisms for processing talker variability. Here, we measured psychophysiological responses to talker variability using electroencephalography (EEG) and pupillometry while listeners performed a delayed recall of digit span task. Listeners heard and recalled seven-digit sequences with both talker (single- vs. mixed-talker digits) and temporal (0- vs. 500-ms inter-digit intervals) discontinuities. Talker discontinuity reduced serial recall accuracy. Both talker and temporal discontinuities elicited P3a-like neural evoked response, while rapid processing of mixed-talkers' speech led to increased phasic pupil dilation. Furthermore, mixed-talkers' speech produced less alpha oscillatory power during working memory maintenance, but not during speech encoding. Overall, these results are consistent with an auditory attention and streaming framework in which talker discontinuity leads to involuntary, stimulus-driven attentional reorientation to novel speech sources, resulting in the processing interference classically associated with talker variability.


Subject(s)
Speech Perception , Speech , Electroencephalography , Humans , Memory, Short-Term , Mental Recall
19.
J Speech Lang Hear Res ; 64(8): 2956-2976, 2021 08 09.
Article in English | MEDLINE | ID: mdl-34297606

ABSTRACT

Purpose We examined how consonant perception is affected by a preceding speech carrier simulated in the same or a different room, for different classes of consonants. Carrier room, carrier length, and carrier length/target room uncertainty were manipulated. A phonetic feature analysis tested which phonetic categories are influenced by the manipulations in the acoustic context of the carrier. Method Two experiments were performed, each with nine participants. Targets consisted of 10 or 16 vowel-consonant (VC) syllables presented in one of two strongly reverberant rooms, preceded by a multiple-VC carrier presented in either the same room, a different reverberant room, or an anechoic room. In Experiment 1, the carrier length and the target room randomly varied from trial to trial, whereas in Experiment 2, they were fixed within a block of trials. Results Overall, a consistent carrier provided an advantage for consonant perception compared to inconsistent carriers, whether in anechoic or differently reverberant rooms. Phonetic analysis showed that carrier inconsistency significantly degraded identification of the manner of articulation, especially for stop consonants and, in one of the rooms, also of voicing. Carrier length and carrier/target uncertainty did not affect adaptation to reverberation for individual phonetic features. The detrimental effects of anechoic and different reverberant carriers on target perception were similar. Conclusions The strength of calibration varies across different phonetic features, as well as across rooms with different levels of reverberation. Even though place of articulation is the feature that is affected by reverberation the most, it is the manner of articulation and, partially, voicing for which room adaptation is observed.


Subject(s)
Speech Perception , Voice , Calibration , Humans , Phonetics , Speech
20.
J Acoust Soc Am ; 149(1): 259, 2021 01.
Article in English | MEDLINE | ID: mdl-33514136

ABSTRACT

The ability to discriminate frequency differences between pure tones declines as the duration of the interstimulus interval (ISI) increases. The conventional explanation for this finding is that pitch representations gradually decay from auditory short-term memory. Gradual decay means that internal noise increases with increasing ISI duration. Another possibility is that pitch representations experience "sudden death," disappearing without a trace from memory. Sudden death means that listeners guess (respond at random) more often when the ISIs are longer. Since internal noise and guessing probabilities influence the shape of psychometric functions in different ways, they can be estimated simultaneously. Eleven amateur musicians performed a two-interval, two-alternative forced-choice frequency-discrimination task. The frequencies of the first tones were roved, and frequency differences and ISI durations were manipulated across trials. Data were analyzed using Bayesian models that simultaneously estimated internal noise and guessing probabilities. On average across listeners, internal noise increased monotonically as a function of increasing ISI duration, suggesting that gradual decay occurred. The guessing rate decreased with an increasing ISI duration between 0.5 and 2 s but then increased with further increases in ISI duration, suggesting that sudden death occurred but perhaps only at longer ISIs. Results are problematic for decay-only models of discrimination and contrast with those from a study on visual short-term memory, which found that over similar durations, visual representations experienced little gradual decay yet substantial sudden death.


Subject(s)
Memory, Short-Term , Music , Pitch Discrimination , Bayes Theorem , Humans , Noise
SELECTION OF CITATIONS
SEARCH DETAIL
...