Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 52
Filter
1.
JASA Express Lett ; 3(12)2023 12 01.
Article in English | MEDLINE | ID: mdl-38038677

ABSTRACT

The effects of sound segregation cues on the sensitivity to intensity increments were explored. Listeners indicated whether the second and fourth sounds (harmonic complexes) within a five-sound sequence were increased in intensity. The target sound had a fundamental frequency of 250 Hz. In different conditions, nontarget sounds had different fundamental frequencies, different spectral shapes, and unique frequency regions relative to the target. For targets more intense than nontargets, nontarget characteristics did not affect thresholds. For targets less intense than the nontargets, thresholds improved when the targets and nontargets had unique frequency regions.


Subject(s)
Cues , Social Segregation , Auditory Perception , Sound , Discrimination, Psychological
2.
J Acoust Soc Am ; 151(5): 3116, 2022 05.
Article in English | MEDLINE | ID: mdl-35649891

ABSTRACT

Acoustics research involving human participants typically takes place in specialized laboratory settings. Listening studies, for example, may present controlled sounds using calibrated transducers in sound-attenuating or anechoic chambers. In contrast, remote testing takes place outside of the laboratory in everyday settings (e.g., participants' homes). Remote testing could provide greater access to participants, larger sample sizes, and opportunities to characterize performance in typical listening environments at the cost of reduced control of environmental conditions, less precise calibration, and inconsistency in attentional state and/or response behaviors from relatively smaller sample sizes and unintuitive experimental tasks. The Acoustical Society of America Technical Committee on Psychological and Physiological Acoustics launched the Task Force on Remote Testing (https://tcppasa.org/remotetesting/) in May 2020 with goals of surveying approaches and platforms available to support remote testing and identifying challenges and considerations for prospective investigators. The results of this task force survey were made available online in the form of a set of Wiki pages and summarized in this report. This report outlines the state-of-the-art of remote testing in auditory-related research as of August 2021, which is based on the Wiki and a literature search of papers published in this area since 2020, and provides three case studies to demonstrate feasibility during practice.


Subject(s)
Acoustics , Auditory Perception , Attention/physiology , Humans , Prospective Studies , Sound
3.
J Acoust Soc Am ; 150(4): 2327, 2021 10.
Article in English | MEDLINE | ID: mdl-34717459

ABSTRACT

Previous studies of level discrimination reported that listeners with high-frequency sensorineural hearing loss (SNHL) place greater weight on high frequencies than normal-hearing (NH) listeners. It is not clear whether these results are influenced by stimulus factors (e.g., group differences in presentation levels, cross-frequency discriminability of level differences used to measure weights) and whether such weights generalize to other tasks. Here, NH and SNHL weights were measured for level, duration, and frequency discrimination of two-tone complexes after measuring discriminability just-noticeable differences for each frequency and stimulus dimension. Stimuli were presented at equal sensation level (SL) or equal sound pressure level (SPL). Results showed that weights could change depending on which frequency contained the more discriminable level difference with uncontrolled cross-frequency discriminability. When cross-frequency discriminability was controlled, weights were consistent for level and duration discrimination, but not for frequency discrimination. Comparing equal SL and equal SPL weights indicated greater weight on the higher-level tone for level and duration discrimination. Weights were unrelated to improvements in recognition of low-pass-filtered speech with increasing cutoff frequency. These results suggest that cross-frequency weights and NH and SNHL weighting differences are influenced by stimulus factors and may not generalize to the use of speech cues in specific frequency regions.


Subject(s)
Hearing Loss, Sensorineural , Speech Perception , Auditory Threshold , Cues , Hearing , Hearing Loss, Sensorineural/diagnosis , Humans , Speech
4.
Hear Res ; 408: 108307, 2021 09 01.
Article in English | MEDLINE | ID: mdl-34311190

ABSTRACT

We recently developed a method to estimate speech-driven spectrotemporal receptive fields (STRFs) using fMRI. The method uses spectrotemporal modulation filtering, a form of acoustic distortion that renders speech sometimes intelligible and sometimes unintelligible. Using this method, we found significant STRF responses only in classic auditory regions throughout the superior temporal lobes. However, our analysis was not optimized to detect small clusters of STRFs as might be expected in non-auditory regions. Here, we re-analyze our data using a more sensitive multivariate statistical test for cross-subject alignment of STRFs, and we identify STRF responses in non-auditory regions including the left dorsal premotor cortex (dPM), left inferior frontal gyrus (IFG), and bilateral calcarine sulcus (calcS). All three regions responded more to intelligible than unintelligible speech, but left dPM and calcS responded significantly to vocal pitch and demonstrated strong functional connectivity with early auditory regions. Left dPM's STRF generated the best predictions of activation on trials rated as unintelligible by listeners, a hallmark auditory profile. IFG, on the other hand, responded almost exclusively to intelligible speech and was functionally connected with classic speech-language regions in the superior temporal sulcus and middle temporal gyrus. IFG's STRF was also (weakly) able to predict activation on unintelligible trials, suggesting the presence of a partial 'acoustic trace' in the region. We conclude that left dPM is part of the human dorsal laryngeal motor cortex, a region previously shown to be capable of operating in an 'auditory mode' to encode vocal pitch. Further, given previous observations that IFG is involved in syntactic working memory and/or processing of linear order, we conclude that IFG is part of a higher-order speech circuit that exerts a top-down influence on processing of speech acoustics. Finally, because calcS is modulated by emotion, we speculate that changes in the quality of vocal pitch may have contributed to its response.


Subject(s)
Auditory Cortex , Speech Perception , Acoustic Stimulation , Auditory Cortex/diagnostic imaging , Brain Mapping , Humans , Magnetic Resonance Imaging , Speech
5.
J Acoust Soc Am ; 149(3): 1567, 2021 03.
Article in English | MEDLINE | ID: mdl-33765831

ABSTRACT

When spectral components of a complex sound are presented not simultaneously but distributed over time, human listeners can still, to a degree, perceptually recover the spectral profile of the sound. This capability of integrating spectral information over time was investigated using a cued informational masking paradigm. Listeners detected a 1-kHz pure tone in a simultaneous masker composed of six random-frequency tones drawn on every trial. The spectral profile of the masker was cued using a precursor sound that consisted of a sequence of 50-ms bursts, separated by inter-burst intervals of 100 ms. Each burst in the precursor consisted of pure tones at the masker frequencies with tones appearing at each of the masker frequencies at different presentation probabilities. As the presentation probability increased in different conditions, the detectability of the target improved, indicating reliable precursor cuing regarding the spectral content of the masker. For many listeners, performance did not significantly improve as the number of precursor bursts increased from 2 to 16, indicating inefficient integration of information beyond 2 bursts. Additional analyses suggest that when intensity of the bursts is relatively constant, the contribution of the precursor is dominated by information in the initial burst.


Subject(s)
Perceptual Masking , Sound , Auditory Threshold , Cues , Humans
6.
J Speech Lang Hear Res ; 63(11): 3816-3833, 2020 11 13.
Article in English | MEDLINE | ID: mdl-33049147

ABSTRACT

Purpose Spatially separating speech and background noise improves speech understanding in normal-hearing listeners, an effect referred to as spatial release from masking (SRM). In cochlear implant (CI) users, SRM has often been demonstrated using asymmetric noise configurations, which maximize benefit from head shadow and the potential availability of binaural cues. In contrast, SRM in symmetrical configurations has been minimal to absent in CI users. We examined the interaction between two types of maskers (informational and energetic) and SRM in bimodal and bilateral CI users. We hypothesized that SRM would be absent or "negative" using symmetrically separated noise maskers. Second, we hypothesized that bimodal listeners would exhibit greater release from informational masking due to access to acoustic information in the non-CI ear. Method Participants included 10 bimodal and 10 bilateral CI users. Speech understanding in noise was tested in 24 conditions: 3 spatial configurations (S0N0, S0N45&315, S0N90&270) × 2 masker types (speech, signal-correlated noise) × 2 listening configurations (best-aided, CI-alone) × 2 talker gender conditions (different-gender, same-gender). Results In support of our first hypothesis, both groups exhibited negative SRM with increasing spatial separation. In opposition to our second hypothesis, both groups exhibited similar magnitudes of release from informational masking. The magnitude of release was greater for bimodal listeners, though this difference failed to reach statistical significance. Conclusions Both bimodal and bilateral CI recipients exhibited negative SRM. This finding is consistent with CI signal processing limitations, the audiologic factors associated with SRM, and known effects of behind-the-ear microphone technology. Though release from informational masking was not significantly different across groups, the magnitude of release was greater for bimodal listeners. This suggests that bimodal listeners may be at least marginally more susceptible to informational masking than bilateral CI users, though further research is warranted.


Subject(s)
Cochlear Implantation , Cochlear Implants , Speech Perception , Humans , Noise , Perceptual Masking
7.
J Acoust Soc Am ; 147(5): 3523, 2020 05.
Article in English | MEDLINE | ID: mdl-32486827

ABSTRACT

Results of simultaneous notched-noise masking are commonly interpreted as reflecting the bandwidth of underlying auditory filters. This interpretation assumes that listeners detect a tone added to notched-noise based on an increase in energy at the output of an auditory filter. Previous work challenged this assumption by showing that randomly and independently varying (roving) the levels of each stimulus interval does not substantially worsen listener thresholds [Lentz, Richards, and Matiasek (1999). J. Acoust. Soc. Am. 106, 2779-2792]. Lentz et al. further challenged this assumption by showing that filter bandwidths based on notched-noise results were different from those based on a profile-analysis task [Green (1983). Am. Psychol. 38, 133-142; (1988). (Oxford University Press, New York)], although these estimates were later reconciled by emphasizing spectral peaks of the profile-analysis stimulus [Lentz (2006). J. Acoust. Soc. Am. 120, 945-956]. Here, a single physiological model is shown to account for performance in fixed- and roving-level notched-noise tasks and the Lentz et al. profile-analysis task. This model depends on peripheral neural fluctuation cues that are transformed into the average rates of model inferior colliculus neurons. Neural fluctuations are influenced by peripheral filters, synaptic adaptation, cochlear amplification, and saturation of inner hair cells, an element not included in previous theories of envelope-based cues for these tasks. Results suggest reevaluation of the interpretation of performance in these paradigms.


Subject(s)
Cues , Perceptual Masking , Auditory Threshold , Mesencephalon , Noise/adverse effects
8.
Front Neurosci ; 13: 999, 2019.
Article in English | MEDLINE | ID: mdl-31607846

ABSTRACT

OBJECTIVES: This study examined musical sound quality (SQ) in adult cochlear implant (CI) recipients. The study goals were to determine: the number of channels needed for high levels of musical SQ overall and by musical genre; the impact of device and patient factors on musical SQ ratings; and the relationship between musical SQ, speech recognition, and speech SQ to relate these findings to measures frequently used in clinical protocols. METHODS: Twenty-one post-lingually deafened adult CI recipients participated in this study. Electrode placement, including scalar location, average electrode-to-modiolus distance ( M ¯ ), and angular insertion depth were determined by CT imaging using validated CI position analysis algorithms (e.g., Noble et al., 2013; Zhao et al., 2018, 2019). CI programs were created using 4-22 electrodes with equal spatial distribution of active electrodes across the array. Speech recognition, speech SQ, music perception via a frequency discrimination task, and musical SQ were acutely assessed for all electrode conditions. Musical SQ was assessed using pre-selected musical excerpts from a variety of musical genres. RESULTS: CI recipients demonstrated continuous improvement in qualitative judgments of musical SQ with up to 10 active electrodes. Participants with straight electrodes placed in scala tympani (ST) and pre-curved electrodes with higher M ¯ variance reported higher levels of musical SQ; however, this relationship is believed to be driven by levels of musical experience as well as the potential for preoperative bias in device selection. Participants reported significant increases in musical SQ beyond four channels for all musical genres examined in the current study except for Hip Hop/Rap. After musical experience outliers were removed, there was no relationship between musical experience or frequency discrimination ability and musical SQ ratings. There was a weak, but significant correlation between qualitative ratings for speech stimuli presented in quiet and in noise and musical SQ. CONCLUSION: Modern CI recipients may need more channels for musical SQ than even required for asymptotic speech recognition or speech SQ. These findings may be used to provide clinical guidance for personalized expectations management of music appreciation depending on individual device and patient factors.

9.
J Acoust Soc Am ; 145(5): EL442, 2019 05.
Article in English | MEDLINE | ID: mdl-31153351

ABSTRACT

To evaluate the ability of a restricted range of auditory-nerve fibers to encode a large perceptual dynamic range, Viemeister [(1983). Science 221, 1206-1208] examined the detection of a change in the level of a high-frequency band of noise flanked by more intense fixed-level noise maskers. Here, stimuli and procedures similar to Viemeister's were used, but random manipulations of level and notch cutoff frequency were included to evaluate predictions of energy-based models. The results indicate that cues other than the change in level per se are available, and suggest the potential contribution of changes in pitch/timbre for this task.


Subject(s)
Auditory Perception/physiology , Cues , Noise , Perceptual Masking/physiology , Pitch Discrimination/physiology , Acoustic Stimulation/methods , Adult , Auditory Threshold/physiology , Female , Humans , Male , Young Adult
10.
Neuroimage ; 200: 490-500, 2019 10 15.
Article in English | MEDLINE | ID: mdl-31254649

ABSTRACT

Natural speech is organized according to a hierarchical structure, with individual speech sounds combining to form abstract linguistic units, and abstract linguistic units combining to form higher-order linguistic units. Since the boundaries between these units are not always indicated by acoustic cues, they must often be computed internally. Signatures of this internal computation were reported by Ding et al. (2016), who presented isochronous sequences of mono-syllabic words that combined to form phrases that combined to form sentences, and showed that cortical responses simultaneously encode boundaries at multiple levels of the linguistic hierarchy. In the present study, we designed melodic sequences that were hierarchically organized according to Western music conventions. Specifically, isochronous sequences of "sung" nonsense syllables were constructed such that syllables combined to form triads outlining individual chords, which combined to form harmonic progressions. EEG recordings were made while participants listened to these sequences with the instruction to detect when violations in the sequence structure occurred. We show that cortical responses simultaneously encode boundaries at multiple levels of a melodic hierarchy, suggesting that the encoding of hierarchical structure is not unique to speech. No effect of musical training on cortical encoding was observed.


Subject(s)
Auditory Perception/physiology , Cerebral Cortex/physiology , Functional Neuroimaging , Music , Adolescent , Adult , Electroencephalography , Female , Humans , Male , Middle Aged , Speech Perception/physiology , Young Adult
11.
J Speech Lang Hear Res ; 62(4): 1051-1067, 2019 04 15.
Article in English | MEDLINE | ID: mdl-30986140

ABSTRACT

Purpose Age-related sensorineural hearing loss can dramatically affect speech recognition performance due to reduced audibility and suprathreshold distortion of spectrotemporal information. Normal aging produces changes within the central auditory system that impose further distortions. The goal of this study was to characterize the effects of aging and hearing loss on perceptual representations of speech. Method We asked whether speech intelligibility is supported by different patterns of spectrotemporal modulations (STMs) in older listeners compared to young normal-hearing listeners. We recruited 3 groups of participants: 20 older hearing-impaired (OHI) listeners, 19 age-matched normal-hearing listeners, and 10 young normal-hearing (YNH) listeners. Listeners performed a speech recognition task in which randomly selected regions of the speech STM spectrum were revealed from trial to trial. The overall amount of STM information was varied using an up-down staircase to hold performance at 50% correct. Ordinal regression was used to estimate weights showing which regions of the STM spectrum were associated with good performance (a "classification image" or CImg). Results The results indicated that (a) large-scale CImg patterns did not differ between the 3 groups; (b) weights in a small region of the CImg decreased systematically as hearing loss increased; (c) CImgs were also nonsystematically distorted in OHI listeners, and the magnitude of this distortion predicted speech recognition performance even after accounting for audibility; and (d) YNH listeners performed better overall than the older groups. Conclusion We conclude that OHI/older normal-hearing listeners rely on the same speech STMs as YNH listeners but encode this information less efficiently. Supplemental Material https://doi.org/10.23641/asha.7859981.


Subject(s)
Age Factors , Hearing Loss, Sensorineural/psychology , Hearing/physiology , Speech Intelligibility/physiology , Speech Perception/physiology , Adolescent , Adult , Aged , Audiometry, Pure-Tone , Auditory Threshold , Case-Control Studies , Female , Humans , Male , Middle Aged , Sound Spectrography , Young Adult
12.
J Speech Lang Hear Res ; 62(2): 442-455, 2019 02 26.
Article in English | MEDLINE | ID: mdl-30950687

ABSTRACT

Purpose A Bayesian adaptive procedure, that is, the quick auditory filter (qAF) procedure, has been shown to improve the efficiency for estimating auditory filter shapes of listeners with normal hearing. The current study evaluates the accuracy and test-retest reliability of the qAF procedure for naïve listeners with a variety of ages and hearing status. Method Fifty listeners who were naïve to psychophysical experiments and exhibit wide ranges of age (19-70 years) and hearing threshold (-5 to 70 dB HL at 2 kHz) were recruited. Their auditory filter shapes were estimated for a 15-dB SL target tone at 2 kHz using both the qAF procedure and the traditional threshold-based procedure. The auditory filter model was defined using 3 parameters: (a) the sharpness of the tip portion of the auditory filter, p; (b) the prominence of the low-frequency tail of the filter, 10log( w); and (c) the listener's efficiency in detection, 10log( K). Results The estimated parameters of the auditory filter model were consistent between 2 qAF runs tested on 2 separate days. The parameter estimates from the 2 qAF runs also agreed well with those estimated using the traditional procedure despite being substantially faster. Across the 3 auditory filter estimates, the dependence of the auditory filter parameters on listener age and hearing threshold was consistent across procedures, as well as consistent with previously published estimates. Conclusions The qAF procedure demonstrates satisfactory test-retest reliability and good agreement to the traditional procedure for listeners with a wide range of ages and with hearing status ranging from normal hearing to moderate hearing impairment.


Subject(s)
Auditory Perception/physiology , Hearing/physiology , Noise , Perceptual Masking/physiology , Adult , Aged , Auditory Threshold/physiology , Bayes Theorem , Discrimination, Psychological/physiology , Humans , Middle Aged , Reproducibility of Results , Young Adult
13.
Neuroimage ; 186: 647-666, 2019 02 01.
Article in English | MEDLINE | ID: mdl-30500424

ABSTRACT

Existing data indicate that cortical speech processing is hierarchically organized. Numerous studies have shown that early auditory areas encode fine acoustic details while later areas encode abstracted speech patterns. However, it remains unclear precisely what speech information is encoded across these hierarchical levels. Estimation of speech-driven spectrotemporal receptive fields (STRFs) provides a means to explore cortical speech processing in terms of acoustic or linguistic information associated with characteristic spectrotemporal patterns. Here, we estimate STRFs from cortical responses to continuous speech in fMRI. Using a novel approach based on filtering randomly-selected spectrotemporal modulations (STMs) from aurally-presented sentences, STRFs were estimated for a group of listeners and categorized using a data-driven clustering algorithm. 'Behavioral STRFs' highlighting STMs crucial for speech recognition were derived from intelligibility judgments. Clustering revealed that STRFs in the supratemporal plane represented a broad range of STMs, while STRFs in the lateral temporal lobe represented circumscribed STM patterns important to intelligibility. Detailed analysis recovered a bilateral organization with posterior-lateral regions preferentially processing STMs associated with phonological information and anterior-lateral regions preferentially processing STMs associated with word- and phrase-level information. Regions in lateral Heschl's gyrus preferentially processed STMs associated with vocalic information (pitch).


Subject(s)
Auditory Cortex/physiology , Brain Mapping/methods , Language , Speech Intelligibility/physiology , Speech Perception/physiology , Adult , Auditory Cortex/diagnostic imaging , Female , Humans , Magnetic Resonance Imaging , Male , Young Adult
14.
J Neurophysiol ; 118(6): 3144-3151, 2017 12 01.
Article in English | MEDLINE | ID: mdl-28877963

ABSTRACT

It has been suggested that cortical entrainment plays an important role in speech perception by helping to parse the acoustic stimulus into discrete linguistic units. However, the question of whether the entrainment response to speech depends on the intelligibility of the stimulus remains open. Studies addressing this question of intelligibility have, for the most part, significantly distorted the acoustic properties of the stimulus to degrade the intelligibility of the speech stimulus, making it difficult to compare across "intelligible" and "unintelligible" conditions. To avoid these acoustic confounds, we used priming to manipulate the intelligibility of vocoded speech. We used EEG to measure the entrainment response to vocoded target sentences that are preceded by natural speech (nonvocoded) prime sentences that are either valid (match the target) or invalid (do not match the target). For unintelligible speech, valid primes have the effect of restoring intelligibility. We compared the effect of priming on the entrainment response for both 3-channel (unintelligible) and 16-channel (intelligible) speech. We observed a main effect of priming, suggesting that the entrainment response depends on prior knowledge, but not a main effect of vocoding (16 channels vs. 3 channels). Furthermore, we found no difference in the effect of priming on the entrainment response to 3-channel and 16-channel vocoded speech, suggesting that for vocoded speech, entrainment response does not depend on intelligibility.NEW & NOTEWORTHY Neural oscillations have been implicated in the parsing of speech into discrete, hierarchically organized units. Our data suggest that these oscillations track the acoustic envelope rather than more abstract linguistic properties of the speech stimulus. Our data also suggest that prior experience with the stimulus allows these oscillations to better track the stimulus envelope.


Subject(s)
Cerebral Cortex/physiology , Repetition Priming , Speech Intelligibility , Speech Perception , Adult , Female , Humans , Male
15.
J Acoust Soc Am ; 140(2): 1072, 2016 08.
Article in English | MEDLINE | ID: mdl-27586738

ABSTRACT

Speech intelligibility depends on the integrity of spectrotemporal patterns in the signal. The current study is concerned with the speech modulation power spectrum (MPS), which is a two-dimensional representation of energy at different combinations of temporal and spectral (i.e., spectrotemporal) modulation rates. A psychophysical procedure was developed to identify the regions of the MPS that contribute to successful reception of auditory sentences. The procedure, based on the two-dimensional image classification technique known as "bubbles" (Gosselin and Schyns (2001). Vision Res. 41, 2261-2271), involves filtering (i.e., degrading) the speech signal by removing parts of the MPS at random, and relating filter patterns to observer performance (keywords identified) over a number of trials. The result is a classification image (CImg) or "perceptual map" that emphasizes regions of the MPS essential for speech intelligibility. This procedure was tested using normal-rate and 2×-time-compressed sentences. The results indicated: (a) CImgs could be reliably estimated in individual listeners in relatively few trials, (b) CImgs tracked changes in spectrotemporal modulation energy induced by time compression, though not completely, indicating that "perceptual maps" deviated from physical stimulus energy, and


Subject(s)
Speech Intelligibility/physiology , Adult , Auditory Threshold , Female , Humans , Male , Psychoacoustics , Speech Perception
16.
Brain Res ; 1644: 203-12, 2016 08 01.
Article in English | MEDLINE | ID: mdl-27195825

ABSTRACT

Recent studies have uncovered a neural response that appears to track the envelope of speech, and have shown that this tracking process is mediated by attention. It has been argued that this tracking reflects a process of phase-locking to the fluctuations of stimulus energy, ensuring that this energy arrives during periods of high neuronal excitability. Because all acoustic stimuli are decomposed into spectral channels at the cochlea, and this spectral decomposition is maintained along the ascending auditory pathway and into auditory cortex, we hypothesized that the overall stimulus envelope is not as relevant to cortical processing as the individual frequency channels; attention may be mediating envelope tracking differentially across these spectral channels. To test this we reanalyzed data reported by Horton et al. (2013), where high-density EEG was recorded while adults attended to one of two competing naturalistic speech streams. In order to simulate cochlear filtering, the stimuli were passed through a gammatone filterbank, and temporal envelopes were extracted at each filter output. Following Horton et al. (2013), the attended and unattended envelopes were cross-correlated with the EEG, and local maxima were extracted at three different latency ranges corresponding to distinct peaks in the cross-correlation function (N1, P2, and N2). We found that the ratio between the attended and unattended cross-correlation functions varied across frequency channels in the N1 latency range, consistent with the hypothesis that attention differentially modulates envelope-tracking activity across spectral channels.


Subject(s)
Attention/physiology , Cerebral Cortex/physiology , Speech Perception/physiology , Acoustic Stimulation , Adult , Auditory Pathways/physiology , Electroencephalography , Evoked Potentials , Evoked Potentials, Auditory , Female , Humans , Male , Signal Processing, Computer-Assisted , Speech Acoustics , Young Adult
17.
J Acoust Soc Am ; 138(6): 3613-24, 2015 Dec.
Article in English | MEDLINE | ID: mdl-26723318

ABSTRACT

Listeners' speech reception is better when speech is masked by a modulated masker compared to an unmodulated masker with the same long-term root-mean-square level. It has been suggested that listeners take advantage of brief periods of quiescence in a modulated masker to extract speech information. Two experiments examined the contribution of such "dip-listening" models. The first experiment estimated psychometric functions for speech intelligibility using sentences masked by sinusoidally modulated and unmodulated speech-shaped noises and the second experiment estimated detection thresholds for a tone pip added at the central dip in the masker. Modulation rates ranging from 1 to 64 Hz were tested. In experiment 1 the slopes of the psychometric functions were shallower for lower modulation rates and the pattern of speech reception thresholds as a function of modulation rate was nonmonotonic with a minimum near 16 Hz. In contrast, the detection thresholds from experiment 2 increased monotonically with modulation rate. The results suggest that the benefits of listening to speech in temporally fluctuating maskers cannot be solely ascribed to the temporal acuity of the auditory system.


Subject(s)
Noise/adverse effects , Perceptual Masking , Psychoacoustics , Recognition, Psychology , Speech Acoustics , Speech Intelligibility , Speech Perception , Voice Quality , Acoustic Stimulation , Adolescent , Adult , Audiometry, Pure-Tone , Auditory Threshold , Cues , Female , Humans , Male , Speech Reception Threshold Test , Young Adult
18.
Behav Res Methods ; 47(1): 13-26, 2015 Mar.
Article in English | MEDLINE | ID: mdl-24671826

ABSTRACT

A MATLAB toolbox for the efficient estimation of the threshold, slope, and lapse rate of the psychometric function is described. The toolbox enables the efficient implementation of the updated maximum-likelihood (UML) procedure. The toolbox uses an object-oriented architecture for organizing the experimental variables and computational algorithms, which provides experimenters with flexibility in experimental design and data management. Descriptions of the UML procedure and the UML Toolbox are provided, followed by toolbox use examples. Finally, guidelines and recommendations of parameter configurations are given.


Subject(s)
Algorithms , Behavioral Research , Electronic Data Processing/methods , Psychometrics , Research Design , Behavioral Research/instrumentation , Behavioral Research/methods , Humans , Psychometrics/instrumentation , Psychometrics/methods
19.
J Acoust Soc Am ; 136(4): 1857-68, 2014 Oct.
Article in English | MEDLINE | ID: mdl-25324086

ABSTRACT

A Bayesian adaptive procedure, the quick-auditory-filter (qAF) procedure, was used to estimate auditory-filter shapes that were asymmetric about their peaks. In three experiments, listeners who were naive to psychoacoustic experiments detected a fixed-level, pure-tone target presented with a spectrally notched noise masker. The qAF procedure adaptively manipulated the masker spectrum level and the position of the masker notch, which was optimized for the efficient estimation of the five parameters of an auditory-filter model. Experiment I demonstrated that the qAF procedure provided a convergent estimate of the auditory-filter shape at 2 kHz within 150 to 200 trials (approximately 15 min to complete) and, for a majority of listeners, excellent test-retest reliability. In experiment II, asymmetric auditory filters were estimated for target frequencies of 1 and 4 kHz and target levels of 30 and 50 dB sound pressure level. The estimated filter shapes were generally consistent with published norms, especially at the low target level. It is known that the auditory-filter estimates are narrower for forward masking than simultaneous masking due to peripheral suppression, a result replicated in experiment III using fewer than 200 qAF trials.


Subject(s)
Auditory Perception , Noise/adverse effects , Perceptual Masking , Acoustic Stimulation , Acoustics , Adolescent , Adult , Audiometry, Pure-Tone , Auditory Threshold , Bayes Theorem , Humans , Pressure , Psychoacoustics , Signal Detection, Psychological , Sound Spectrography , Young Adult
20.
J Acoust Soc Am ; 134(2): 1134-45, 2013 Aug.
Article in English | MEDLINE | ID: mdl-23927113

ABSTRACT

A Bayesian adaptive procedure for estimating the auditory-filter shape was proposed and evaluated using young, normal-hearing listeners at moderate stimulus levels. The resulting quick-auditory-filter (qAF) procedure assumed the power spectrum model of masking with the auditory-filter shape being modeled using a spectrally symmetric, two-parameter rounded-exponential (roex) function. During data collection using the qAF procedure, listeners detected the presence of a pure-tone signal presented in the spectral notch of a noise masker. Dependent on the listener's response on each trial, the posterior probability distributions of the model parameters were updated, and the resulting parameter estimates were then used to optimize the choice of stimulus parameters for the subsequent trials. Results showed that the qAF procedure gave similar parameter estimates to the traditional threshold-based procedure in many cases and was able to reasonably predict the masked signal thresholds. Additional measurements suggested that occasional failures of the qAF procedure to reliably converge could be a consequence of incorrect responses early in a qAF track. The addition of a parameter describing lapses of attention reduced the likelihood of such failures.


Subject(s)
Auditory Pathways/physiology , Auditory Perception , Models, Psychological , Acoustic Stimulation , Adult , Attention , Audiometry, Pure-Tone , Auditory Threshold , Bayes Theorem , Female , Humans , Male , Noise/adverse effects , Perceptual Masking , Pressure , Psychoacoustics , Signal Detection, Psychological , Sound Spectrography , Young Adult
SELECTION OF CITATIONS
SEARCH DETAIL
...