Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 466
Filter
1.
Clin Linguist Phon ; : 1-16, 2024 Jun 04.
Article in English | MEDLINE | ID: mdl-38832412

ABSTRACT

Cleft type affects speech outcomes, but exact relationships remain unclear as outcome measures vary. The primary aim was to investigate the relationship between cleft type and speech outcome using different measures in 4-to-6-year-olds with non-syndromic clefts. Secondary aims were to explore the relationships between (i) speech measures used; and (ii) parent perception of speech intelligibility and listener familiarity. Twenty-two pre-schoolers with clefts, plus one parent for each child, were recruited through a hospital outpatient clinic. Children with cleft lip and palate (CLP; n = 11) and those with cleft palate only (CP; n = 11), matched on age and time of palate repair, were compared on Percentage Consonants Correct (PCC), clinician-reported speech intelligibility, and parent rating on the Intelligibility-in-Context Scale (ICS). Children with CLP had significantly lower PCC scores than children with CP (p = .020), but had no significant differences in their clinician- or parent-reported speech intelligibility. Clinician-reported speech intelligibility correlated significantly with both PCC (τ = .594, p < 0.01) and ICS (τ = .424, p = 0.009). No significant correlation was found between PCC and ICS (τ =.197, p = 0.113). Overall, parents rated their child's intelligibility higher for familiar compared to unfamiliar communication partners (τ = 2.325, p = 0.001, r = .76). Cleft type is crucial for intervention planning when objective measures are employed. Speech outcomes should be evaluated at impairment, activity, and participation levels, and by different communication partners, to comprehensively evaluate communicative effectiveness.

2.
Int J Audiol ; : 1-8, 2024 Jun 04.
Article in English | MEDLINE | ID: mdl-38832702

ABSTRACT

OBJECTIVE: The dynamic range (DR) available to the patient is a central parameter to determine speech intelligibility in quiet. DESIGN: In this retrospective study, the DR for the Vibrant Soundbridge implanted in individual patients was calculated using in situ thresholds of the patients and technical data of the implant system. The average DR across frequencies (0.5, 1, 2, 4 kHz) was correlated with the patients' assigned word recognition score (WRS) in quiet. STUDY SAMPLE: A data set of 66 cases (4 bilateral and 2 revised cases) from 60 implanted patients between 14.3-81.8 years were analysed. RESULTS: The relationship between DR and WRS was described by a sigmoidal growth function with R2=0.6371 and a maximum WRS (upper asymptote) of 93.5%. Word recognition scores in quiet improved with increasing DR. A significant shift in performance was detected from DR bin 2 (10-20 dB, median WRS 55%) to bin 3 (20-30 dB, median WRS 80%) and from DR bin 4 (30-40 dB, median WRS 82.5%) to bin 5 (40-50 dB, median WRS 90%). CONCLUSION: A minimum DR of 20 dB can yield sufficient speech intelligibility in quiet in implanted patients, however, an optimum DR is suggested to be 40 dB.

3.
J Oral Biol Craniofac Res ; 14(4): 455-460, 2024.
Article in English | MEDLINE | ID: mdl-38868459

ABSTRACT

Introduction: Orthognathic surgery results in the positional change of the maxilla and mandible that may affect speech. The present study evaluated the effect of combined maxillary advancement and mandibular setback surgery on articulation proficiency and speech intelligibility in patients with non-syndromic skeletal Class III malocclusion. Methods: In this prospective study, twenty-five patients with skeletal class III malocclusion and consecutively treated with Lefort-1 maxillary advancement and mandibular setback (BSSO) orthognathic surgery were included in this study. The speech sample was recorded with a digital audio tape recorder one day before surgery and at 3, 6, 9, 12 and 18 months after surgery. Three qualified and experienced speech and language pathologists evaluated articulation errors and intelligibility of speech samples. Repeated One-way analysis of variance was used to compare articulation proficiency and speech intelligibility at different time intervals. Results: The substitution, omission, distortion and addition errors showed no significant changes at 3 months and 6 months. The total articulation errors decreased to zero at 9 months and no significant increase was observed till 18 months (P < 0.05). Speech intelligibility showed statistically non-significant improvement at any time interval. Cephalometric skeletal parameters SNA and N l A°. were significantly correlated with addition and total articulation errors at 18 months follow up. Conclusions: The ortho-surgical treatment improves speech (decreases. articulation errors) in most of the patients usually 6-9 months post-surgery. Speech intelligibility is not affected by bimaxillary orthognathic surgery in skeletal class III patients. The articulation errors were correlated to changes in position of maxilla.

4.
J Voice ; 2024 May 15.
Article in English | MEDLINE | ID: mdl-38755076

ABSTRACT

OBJECTIVE: To synthesize existing evidence of the relationship between voice parameters and speech intelligibility. METHODS: Following Preferred Reporting Items for Systematic Reviews and Meta-Analysis extension for Scoping Review (PRISMA-ScR) guidelines, 13 databases were searched and a manual search was conducted. A narrative synthesis of methodological quality, study characteristics, participant demographics, voice parameter categorization, and their relationship to speech intelligibility was conducted. A Grading of Recommendations Assessment, Development, and Evaluation (GRADE) assessment was also performed. RESULTS: A total of 5593 studies were retrieved, and 30 eligible studies were included in the final scoping review. The studies were given scores of 10-25 (average 16.93) out of 34 in the methodological quality assessment. Research that analyzed voice parameters related to speech intelligibility, encompassing perceptual, acoustic, and aerodynamic parameters, was included. Validated and nonvalidated perceptual voice assessments showed divergent results regarding the relationship between perceptual parameters and speech intelligibility. The relationship between acoustic parameters and speech intelligibility was found to be complex and the results were inconsistent. The limited research on aerodynamic parameters did not reach a consensus on their relationship with speech intelligibility. Studies in which listeners were not speech-language pathologists (SLPs) far outnumbered those with SLP listeners, and research conducted in English contexts significantly exceeded that in non-English contexts. The GRADE evaluation indicated that the quality of evidence varied from low to moderate. DISCUSSION: The results for the relationship between voice parameters and intelligibility showed significant heterogeneity. Future research should consider age-related voice changes and include diverse age groups. To enhance validity and comparability, it will be necessary to report effect sizes, tool validity, inter-rater reliability, and calibration procedures. Voice assessments should account for the validation status of tools because of their potential impact on the outcomes. The linguistic context may also influence the results.

5.
Front Artif Intell ; 7: 1359094, 2024.
Article in English | MEDLINE | ID: mdl-38800762

ABSTRACT

Perceptual measures, such as intelligibility and speech disorder severity, are widely used in the clinical assessment of speech disorders in patients treated for oral or oropharyngeal cancer. Despite their widespread usage, these measures are known to be subjective and hard to reproduce. Therefore, an M-Health assessment based on an automatic prediction has been seen as a more robust and reliable alternative. Despite recent progress, these automatic approaches still remain somewhat theoretical, and a need to implement them in real clinical practice rises. Hence, in the present work we introduce SAMI, a clinical mobile application used to predict speech intelligibility and disorder severity as well as to monitor patient progress on these measures over time. The first part of this work illustrates the design and development of the systems supported by SAMI. Here, we show how deep neural speaker embeddings are used to automatically regress speech disorder measurements (intelligibility and severity), as well as the training and validation of the system on a French corpus of head and neck cancer. Furthermore, we also test our model on a secondary corpus recorded in real clinical conditions. The second part details the results obtained from the deployment of our system in a real clinical environment, over the course of several weeks. In this section, the results obtained with SAMI are compared to an a posteriori perceptual evaluation, conducted by a set of experts on the new recorded data. The comparison suggests a high correlation and a low error between the perceptual and automatic evaluations, validating the clinical usage of the proposed application.

6.
Audiol Res ; 14(3): 469-478, 2024 May 22.
Article in English | MEDLINE | ID: mdl-38804463

ABSTRACT

The use of remote testing to collect behavioral data has been on the rise, especially after the COVID-19 pandemic. Here we present psychometric functions for a commonly used speech corpus obtained in remote testing and laboratory testing conditions on young normal hearing listeners in the presence of different types of maskers. Headphone use for the remote testing group was checked by supplementing procedures from prior literature using a Huggins pitch task. Results revealed no significant differences in the measured thresholds using the remote testing and laboratory testing conditions for all the three masker types. Also, the thresholds measured obtained in these two conditions were strongly correlated for a different group of young normal hearing listeners. Based on the results, excellent outcomes on auditory threshold measurements where the stimuli are presented both at levels lower than and above an individual's speech-recognition threshold can be obtained by remotely testing the listeners.

7.
Cogn Res Princ Implic ; 9(1): 25, 2024 04 23.
Article in English | MEDLINE | ID: mdl-38652383

ABSTRACT

The use of face coverings can make communication more difficult by removing access to visual cues as well as affecting the physical transmission of speech sounds. This study aimed to assess the independent and combined contributions of visual and auditory cues to impaired communication when using face coverings. In an online task, 150 participants rated videos of natural conversation along three dimensions: (1) how much they could follow, (2) how much effort was required, and (3) the clarity of the speech. Visual and audio variables were independently manipulated in each video, so that the same video could be presented with or without a superimposed surgical-style mask, accompanied by one of four audio conditions (either unfiltered audio, or audio-filtered to simulate the attenuation associated with a surgical mask, an FFP3 mask, or a visor). Hypotheses and analyses were pre-registered. Both the audio and visual variables had a statistically significant negative impact across all three dimensions. Whether or not talkers' faces were visible made the largest contribution to participants' ratings. The study identifies a degree of attenuation whose negative effects can be overcome by the restoration of visual cues. The significant effects observed in this nominally low-demand task (speech in quiet) highlight the importance of the visual and audio cues in everyday life and that their consideration should be included in future face mask designs.


Subject(s)
Cues , Speech Perception , Humans , Adult , Female , Male , Young Adult , Speech Perception/physiology , Visual Perception/physiology , Masks , Adolescent , Speech/physiology , Communication , Middle Aged , Facial Recognition/physiology
8.
Trends Hear ; 28: 23312165241240572, 2024.
Article in English | MEDLINE | ID: mdl-38676325

ABSTRACT

Realistic outcome measures that reflect everyday hearing challenges are needed to assess hearing aid and cochlear implant (CI) fitting. Literature suggests that listening effort measures may be more sensitive to differences between hearing-device settings than established speech intelligibility measures when speech intelligibility is near maximum. Which method provides the most effective measurement of listening effort for this purpose is currently unclear. This study aimed to investigate the feasibility of two tests for measuring changes in listening effort in CI users due to signal-to-noise ratio (SNR) differences, as would arise from different hearing-device settings. By comparing the effect size of SNR differences on listening effort measures with test-retest differences, the study evaluated the suitability of these tests for clinical use. Nineteen CI users underwent two listening effort tests at two SNRs (+4 and +8 dB relative to individuals' 50% speech perception threshold). We employed dual-task paradigms-a sentence-final word identification and recall test (SWIRT) and a sentence verification test (SVT)-to assess listening effort at these two SNRs. Our results show a significant difference in listening effort between the SNRs for both test methods, although the effect size was comparable to the test-retest difference, and the sensitivity was not superior to speech intelligibility measures. Thus, the implementations of SVT and SWIRT used in this study are not suitable for clinical use to measure listening effort differences of this magnitude in individual CI users. However, they can be used in research involving CI users to analyze group data.


Subject(s)
Cochlear Implantation , Cochlear Implants , Feasibility Studies , Persons With Hearing Impairments , Speech Intelligibility , Speech Perception , Humans , Male , Female , Speech Perception/physiology , Middle Aged , Aged , Speech Intelligibility/physiology , Cochlear Implantation/instrumentation , Persons With Hearing Impairments/rehabilitation , Persons With Hearing Impairments/psychology , Reproducibility of Results , Acoustic Stimulation , Signal-To-Noise Ratio , Adult , Aged, 80 and over , Auditory Threshold/physiology , Predictive Value of Tests , Correction of Hearing Impairment/instrumentation , Noise/adverse effects
9.
Front Neurosci ; 18: 1368641, 2024.
Article in English | MEDLINE | ID: mdl-38646607

ABSTRACT

Noise-vocoded speech has long been used to investigate how acoustic cues affect speech understanding. Studies indicate that reducing the number of spectral channel bands diminishes speech intelligibility. Despite previous studies examining the channel band effect using earlier event-related potential (ERP) components, such as P1, N1, and P2, a clear consensus or understanding remains elusive. Given our hypothesis that spectral degradation affects higher-order processing of speech understanding beyond mere perception, we aimed to objectively measure differences in higher-order abilities to discriminate or interpret meaning. Using an oddball paradigm with speech stimuli, we examined how neural signals correlate with the evaluation of speech stimuli based on the number of channel bands measuring N2 and P3b components. In 20 young participants with normal hearing, we measured speech intelligibility and N2 and P3b responses using a one-syllable task paradigm with animal and non-animal stimuli across four vocoder conditions with 4, 8, 16, or 32 channel bands. Behavioral data from word repetition clearly affected the number of channel bands, and all pairs were significantly different (p < 0.001). We also observed significant effects of the number of channels on the peak amplitude [F(2.006, 38.117) = 9.077, p < 0.001] and peak latency [F(3, 57) = 26.642, p < 0.001] of the N2 component. Similarly, the P3b component showed significant main effects of the number of channel bands on the peak amplitude [F(2.231, 42.391) = 13.045, p < 0.001] and peak latency [F(3, 57) = 2.968, p = 0.039]. In summary, our findings provide compelling evidence that spectral channel bands profoundly influence cortical speech processing, as reflected in the N2 and P3b components, a higher-order cognitive process. We conclude that spectrally degraded one-syllable speech primarily affects cortical responses during semantic integration.

10.
Trends Hear ; 28: 23312165241246597, 2024.
Article in English | MEDLINE | ID: mdl-38629486

ABSTRACT

Hearing aids and other hearing devices should provide the user with a benefit, for example, compensate for effects of a hearing loss or cancel undesired sounds. However, wearing hearing devices can also have negative effects on perception, previously demonstrated mostly for spatial hearing, sound quality and the perception of the own voice. When hearing devices are set to transparency, that is, provide no gain and resemble open-ear listening as well as possible, these side effects can be studied in isolation. In the present work, we conducted a series of experiments that are concerned with the effect of transparent hearing devices on speech perception in a collocated speech-in-noise task. In such a situation, listening through a hearing device is not expected to have any negative effect, since both speech and noise undergo identical processing, such that the signal-to-noise ratio at ear is not altered and spatial effects are irrelevant. However, we found a consistent hearing device disadvantage for speech intelligibility and similar trends for rated listening effort. Several hypotheses for the possible origin for this disadvantage were tested by including several different devices, gain settings and stimulus levels. While effects of self-noise and nonlinear distortions were ruled out, the exact reason for a hearing device disadvantage on speech perception is still unclear. However, a significant relation to auditory model predictions demonstrate that the speech intelligibility disadvantage is related to sound quality, and is most probably caused by insufficient equalization, artifacts of frequency-dependent signal processing and processing delays.


Subject(s)
Hearing Aids , Hearing Loss , Speech Perception , Humans , Hearing , Noise/adverse effects
11.
Trends Hear ; 28: 23312165241234202, 2024.
Article in English | MEDLINE | ID: mdl-38549451

ABSTRACT

This study investigates the effect of spatial release from masking (SRM) in bilateral bone conduction (BC) stimulation at the mastoid. Nine adults with normal hearing were tested to determine SRM based on speech recognition thresholds (SRTs) in simulated spatial configurations ranging from 0 to 180 degrees. These configurations were based on nonindividualized head-related transfer functions. The participants were subjected to sound stimulation through either air conduction (AC) via headphones or BC. The results indicated that both the angular separation between the target and the masker, and the modality of sound stimulation, significantly influenced speech recognition performance. As the angular separation between the target and the masker increased up to 150°, both BC and AC SRTs decreased, indicating improved performance. However, performance slightly deteriorated when the angular separation exceeded 150°. For spatial separations less than 75°, BC stimulation provided greater spatial benefits than AC, although this difference was not statistically significant. For separations greater than 75°, AC stimulation offered significantly more spatial benefits than BC. When speech and noise originated from the same side of the head, the "better ear effect" did not significantly contribute to SRM. However, when speech and noise were located on opposite sides of the head, this effect became dominant in SRM.


Subject(s)
Bone Conduction , Speech Perception , Adult , Humans , Mastoid , Perceptual Masking/physiology , Speech Perception/physiology , Hearing
12.
HNO ; 72(7): 504-514, 2024 Jul.
Article in German | MEDLINE | ID: mdl-38536465

ABSTRACT

BACKGROUND: Binaural hearing enables better speech comprehension in noisy environments and is necessary for acoustic spatial orientation. This study investigates speech discrimination in noise with separated signal sources and measures sound localization. The aim was to study characteristics and reproducibility of two selected measurement techniques which seem to be suitable for description of the aforementioned aspects of binaural hearing. MATERIALS AND METHODS: Speech reception thresholds (SRT) in noise and test-retest reliability were collected from 55 normal-hearing adults for a spatial setup of loudspeakers with angles of ±â€¯45° and ±â€¯90° using the Oldenburg sentence test. The investigations of sound localization were conducted in a semicircle and fullcircle setup (7 and 12 equidistant loudspeakers). RESULTS: SRT (S-45N45: -14.1 dB SNR; S45N-45: -16.4 dB SNR; S0N90: -13.1 dB SNR; S0N-90: -13.4 dB SNR) and test-retest reliability (4 to 6 dB SNR) were collected for speech intelligibility in noise with separated signals. The procedural learning effect for this setup could only be mitigated with 120 training sentences. Significantly smaller SRT values, resulting in better speech discrimination, were found for the test situation of the right compared to the left ear. RMS values could be gathered for sound localization in the semicircle (1,9°) as well as in the fullcircle setup (11,1°). Better results were obtained in the retest of the fullcircle setup. CONCLUSION: When using the Oldenburg sentence test in noise with spatially separated signals, it is mandatory to perform a training session of 120 sentences in order to minimize the procedural learning effect. Ear-specific SRT values for speech discrimination in noise with separated signal sources are required, which is probably due to the right-ear advantage. A training is recommended for sound localization in the fullcircle setup.


Subject(s)
Noise , Sound Localization , Speech Perception , Humans , Sound Localization/physiology , Reproducibility of Results , Female , Adult , Male , Speech Perception/physiology , Young Adult , Sensitivity and Specificity , Speech Reception Threshold Test/methods , Acoustic Stimulation/methods , Speech Discrimination Tests/methods
13.
Laryngoscope Investig Otolaryngol ; 9(1): e1204, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38362198

ABSTRACT

Objective: Patients lose their voice after laryngectomy for laryngeal cancer or aspiration prevention surgery for severe dysphagia. To assist such patients, we developed and verified the utility of a novel vocalization method using a device termed the voice retriever (VR), in which the sound source is placed in the mouth. Methods: We investigated the effectiveness of the VR in patients. The VR consists of a mouthpiece with a built-in speaker and a dedicated application that serves as the sound source. We compared the speech intelligibility and naturalness in normal participants using VR and an electrolarynx (EL) for the first time as well as the voice-related quality of life (V-RQOL) in patients with dysphonia before and after using the VR. Results: The VR produced significantly higher 100-syllable test scores as well as fluency, amount of additional noise, intonation, intelligibility and overall long reading test ratings in the first-time VR and EL users. Furthermore, the VR use significantly improved the V-RQOL of participants with dysphonia. Conclusion: Compared to EL, VR allows more effective speech improvement in participants without experience using an alternative vocalization method and improves the V-RQOL in patients with dysphonia. Level of Evidence: Step 4.

14.
Front Hum Neurosci ; 18: 1324027, 2024.
Article in English | MEDLINE | ID: mdl-38410256

ABSTRACT

Introduction: Objectively predicting speech intelligibility is important in both telecommunication and human-machine interaction systems. The classic method relies on signal-to-noise ratios (SNR) to successfully predict speech intelligibility. One exception is clear speech, in which a talker intentionally articulates as if speaking to someone who has hearing loss or is from a different language background. As a result, at the same SNR, clear speech produces higher intelligibility than conversational speech. Despite numerous efforts, no objective metric can successfully predict the clear speech benefit at the sentence level. Methods: We proposed a Syllable-Rate-Adjusted-Modulation (SRAM) index to predict the intelligibility of clear and conversational speech. The SRAM used as short as 1 s speech and estimated its modulation power above the syllable rate. We compared SRAM with three reference metrics: envelope-regression-based speech transmission index (ER-STI), hearing-aid speech perception index version 2 (HASPI-v2) and short-time objective intelligibility (STOI), and five automatic speech recognition systems: Amazon Transcribe, Microsoft Azure Speech-To-Text, Google Speech-To-Text, wav2vec2 and Whisper. Results: SRAM outperformed the three reference metrics (ER-STI, HASPI-v2 and STOI) and the five automatic speech recognition systems. Additionally, we demonstrated the important role of syllable rate in predicting speech intelligibility by comparing SRAM with the total modulation power (TMP) that was not adjusted by the syllable rate. Discussion: SRAM can potentially help understand the characteristics of clear speech, screen speech materials with high intelligibility, and convert conversational speech into clear speech.

15.
Acta Otolaryngol ; 144(1): 44-51, 2024 Jan.
Article in English | MEDLINE | ID: mdl-38400594

ABSTRACT

BACKGROUND: With advances in cochlear implant (CI) technology, prelingual deaf adults may experience improved speech perception and quality of life (QoL). It is still a challenge for Mandarin-speaking CI user with tone recognition due to CI technology focused on intonation language. OBJECTIVES: To evaluate the long-term post-CI auditory performance and social-emotional benefits in prelingual deaf Mandarin-speaking adults and the difference between them and post-lingual deaf adults. MATERIAL AND METHODS: Fifty-five adult implanted ears were included (forty-six postlingual deaf group; nine prelingual deaf group). Post-CI long-term outcomes were using vowels, consonants, disyllabic words, Mandarin monosyllable words, categories of audiology performance, speech intelligibility rating, subjective social-emotional questionnaires. RESULTS: Post-CI auditory performance and speech intelligibility of prelingual deafness adults was significantly inferior to that of those with postlingual deafness. However, both groups presented improved social-emotional benefits, with no significant difference between both groups. CONCLUSIONS: Adult CI recipients who deaf before the age of 4 can experience benefits in social-emotional life functioning, regardless of their limited auditory performance and speech intelligibility. Therefore, prelingual Mandarin-speaking deaf adults, especially those using oral communication, can be considered as relative indications for cochlear implantation. SIGNIFICANCE: To clarify and validate the benefits among Mandarin-speaking prelingual deaf adult recipients.


Subject(s)
Cochlear Implantation , Cochlear Implants , Deafness , Speech Perception , Adult , Humans , Quality of Life , Deafness/surgery , Deafness/rehabilitation
16.
Article in English | MEDLINE | ID: mdl-38237606

ABSTRACT

BACKGROUND: Perceptual measures such as speech intelligibility are known to be biased, variant and subjective, to which an automatic approach has been seen as a more reliable alternative. On the other hand, automatic approaches tend to lack explainability, an aspect that can prevent the widespread usage of these technologies clinically. AIMS: In the present work, we aim to study the relationship between four perceptual parameters and speech intelligibility by automatically modelling the behaviour of six perceptual judges, in the context of head and neck cancer. From this evaluation we want to assess the different levels of relevance of each parameter as well as the different judge profiles that arise, both perceptually and automatically. METHODS AND PROCEDURES: Based on a passage reading task from the Carcinologic Speech Severity Index (C2SI) corpus, six expert listeners assessed the voice quality, resonance, prosody and phonemic distortions, as well as the speech intelligibility of patients treated for oral or oropharyngeal cancer. A statistical analysis and an ensemble of automatic systems, one per judge, were devised, where speech intelligibility is predicted as a function of the four aforementioned perceptual parameters of voice quality, resonance, prosody and phonemic distortions. OUTCOMES AND RESULTS: The results suggest that we can automatically predict speech intelligibility as a function of the four aforementioned perceptual parameters, achieving a high correlation of 0.775 (Spearman's ρ). Furthermore, different judge profiles were found perceptually that were successfully modelled automatically. CONCLUSIONS AND IMPLICATIONS: The four investigated perceptual parameters influence the global rating of speech intelligibility, showing that different judge profiles emerge. The proposed automatic approach displayed a more uniform profile across all judges, displaying a more reliable, unbiased and objective prediction. The system also adds an extra layer of interpretability, since speech intelligibility is regressed as a direct function of the individual prediction of the four perceptual parameters, an improvement over more black box approaches. WHAT THIS PAPER ADDS: What is already known on this subject Speech intelligibility is a clinical measure typically used in the post-treatment assessment of speech affecting disorders, such as head and neck cancer. Their perceptual assessment is currently the main method of evaluation; however, it is known to be quite subjective since intelligibility can be seen as a combination of other perceptual parameters (voice quality, resonance, etc.). Given this, automatic approaches have been seen as a more viable alternative to the traditionally used perceptual assessments. What this study adds to existing knowledge The present work introduces a study based on the relationship between four perceptual parameters (voice quality, resonance, prosody and phonemic distortions) and speech intelligibility, by automatically modelling the behaviour of six perceptual judges. The results suggest that different judge profiles arise, both in the perceptual case as well as in the automatic models. These different profiles found showcase the different schools of thought that perceptual judges have, in comparison to the automatic judges, that display more uniform levels of relevance across all the four perceptual parameters. This aspect shows that an automatic approach promotes unbiased, reliable and more objective predictions. What are the clinical implications of this work? The automatic prediction of speech intelligibility, using a combination of four perceptual parameters, show that these approaches can achieve high correlations with the reference scores while maintaining a certain degree of explainability. The more uniform judge profiles found on the automatic case also display less biased results towards the four perceptual parameters. This aspect facilitates the clinical implementation of this class of systems, as opposed to the more subjective and harder to reproduce perceptual assessments.

17.
Hear Res ; 443: 108949, 2024 Mar 01.
Article in English | MEDLINE | ID: mdl-38281473

ABSTRACT

Spectro-temporal modulation (STM) detection sensitivity has been shown to be associated with speech-in-noise reception in hearing-impaired (HI) individuals. Based on previous research, a recent study [Zaar, Simonsen, Dau, and Laugesen (2023). Hear Res 427:108650] introduced an STM test paradigm with audibility compensation, employing STM stimulus variants using noise and complex tones as carrier signals. The study demonstrated that the test was suitable for the target population of elderly individuals with moderate-to-severe hearing loss and showed promising predictions of speech-reception thresholds (SRTs) measured in a realistic set up with spatially distributed speech and noise maskers and linear audibility compensation. The present study further investigated the suggested STM test with respect to (i) test-retest variability for the most promising STM stimulus variants, (ii) its predictive power with respect to realistic speech-in-noise reception with non-linear hearing-aid amplification, (iii) its connection to effects of directionality and noise reduction (DIR+NR) hearing-aid processing, and (iv) its relation to DIR+NR preference. Thirty elderly HI participants were tested in a combined laboratory and field study, collecting STM thresholds with a complex-tone based and a noise-based STM stimulus design, SRTs with spatially distributed speech and noise maskers using hearing aids with non-linear amplification and two different levels of DIR+NR, as well as subjective reports and preference ratings obtained in two field periods with the two DIR+NR hearing-aid settings. The results indicate that the noise-carrier based STM test variant (i) showed optimal test-retest properties, (ii) yielded a highly significant correlation with SRTs (R2=0.61) exceeding and complementing the predictive power of the audiogram, (iii) yielded significant correlation (R2=0.51) with the DIR+NR-induced SRT benefit, and (iv) did not provide significant correlation with subjective preference for DIR+NR settings in the field. Overall, the suggested STM test represents a valuable tool for diagnosing speech-reception problems that remain when hearing-aid amplification has been provided and the resulting need for and benefit from DIR+NR hearing-aid processing.


Subject(s)
Hearing Aids , Hearing Loss, Sensorineural , Hearing Loss , Speech Perception , Humans , Aged , Speech , Hearing Loss/diagnosis , Hearing , Hearing Loss, Sensorineural/diagnosis
18.
Eur Arch Otorhinolaryngol ; 281(3): 1589-1595, 2024 Mar.
Article in English | MEDLINE | ID: mdl-38175264

ABSTRACT

PURPOSE: Previous studies have shown that levels for 50% speech intelligibility in quiet and in noise differ for different languages. Here, we aimed to find out whether these differences may relate to different auditory processing of temporal sound features in different languages, and to determine the influence of tinnitus on speech comprehension in different languages. METHODS: We measured speech intelligibility under various conditions (words in quiet, sentences in babble noise, interrupted sentences) along with tone detection thresholds in quiet [PTA] and in noise [PTAnoise], gap detection thresholds [GDT], and detection thresholds for frequency modulation [FMT], and compared them between Czech and Swiss subjects matched in mean age and PTA. RESULTS: The Swiss subjects exhibited higher speech reception thresholds in quiet, higher threshold speech-to-noise ratio, and shallower slope of performance-intensity function for the words in quiet. Importantly, the intelligibility of temporally gated speech was similar in the Czech and Swiss subjects. The PTAnoise, GDT, and FMT were similar in the two groups. The Czech subjects exhibited correlations of the speech tests with GDT and FMT, which was not the case in the Swiss group. Qualitatively, the results of comparisons between the Swiss and Czech populations were not influenced by presence of subjective tinnitus. CONCLUSION: The results support the notion of language-specific differences in speech comprehension which persists also in tinnitus subjects, and indicates different associations with the elementary measures of auditory temporal processing.


Subject(s)
Speech Perception , Time Perception , Tinnitus , Humans , Speech Intelligibility , Czech Republic , Switzerland , Auditory Threshold , Perceptual Masking , Auditory Perception , Language
19.
Cortex ; 172: 54-71, 2024 03.
Article in English | MEDLINE | ID: mdl-38215511

ABSTRACT

Cortical tracking of speech is vital for speech segmentation and is linked to speech intelligibility. However, there is no clear consensus as to whether reduced intelligibility leads to a decrease or an increase in cortical speech tracking, warranting further investigation of the factors influencing this relationship. One such factor is listening effort, defined as the cognitive resources necessary for speech comprehension, and reported to have a strong negative correlation with speech intelligibility. Yet, no studies have examined the relationship between speech intelligibility, listening effort, and cortical tracking of speech. The aim of the present study was thus to examine these factors in quiet and distinct adverse listening conditions. Forty-nine normal hearing adults listened to sentences produced casually, presented in quiet and two adverse listening conditions: cafeteria noise and reverberant speech. Electrophysiological responses were registered with electroencephalogram, and listening effort was estimated subjectively using self-reported scores and objectively using pupillometry. Results indicated varying impacts of adverse conditions on intelligibility, listening effort, and cortical tracking of speech, depending on the preservation of the speech temporal envelope. The more distorted envelope in the reverberant condition led to higher listening effort, as reflected in higher subjective scores, increased pupil diameter, and stronger cortical tracking of speech in the delta band. These findings suggest that using measures of listening effort in addition to those of intelligibility is useful for interpreting cortical tracking of speech results. Moreover, reading and phonological skills of participants were positively correlated with listening effort in the cafeteria condition, suggesting a special role of expert language skills in processing speech in this noisy condition. Implications for future research and theories linking atypical cortical tracking of speech and reading disorders are further discussed.


Subject(s)
Listening Effort , Speech Perception , Adult , Humans , Noise , Cognition/physiology , Comprehension , Speech Intelligibility/physiology , Speech Perception/physiology
20.
J Deaf Stud Deaf Educ ; 29(2): 265-277, 2024 Mar 17.
Article in English | MEDLINE | ID: mdl-38079579

ABSTRACT

There is great variability in the ways in which the speech intelligibility of d/Deaf and hard-of-hearing (DHH) children who use spoken language as part, or all, of their communication system is measured. This systematic review examined the measures and methods that have been used when examining the speech intelligibility of children who are DHH and the characteristics of these measures and methods. A systematic database search was conducted of CENTRAL; CINAHL; Cochrane; ERIC; Joanna Briggs; Linguistics, Language and Behavior Abstracts; Medline; Scopus; and Web of Science databases, as well as supplemental searches. A total of 204 included studies reported the use of many different measures/methods which measured segmental aspects of speech, with the most common being Allen et al.'s (2001, The reliability of a rating scale for measuring speech intelligibility following pediatric cochlear implantation. Otology and Neurotology, 22(5), 631-633. https://doi.org/10.1097/00129492-200109000-00012) Speech Intelligibility Rating scale. Many studies included insufficient details to determine the measure that was used. Future research should utilize methods/measures with known psychometric validity, provide clear descriptions of the methods/measures used, and consider using more than one measure to account for limitations inherent in different methods of measuring the speech intelligibility of children who are DHH, and consider and discuss the rationale for the measure/method chosen.


Subject(s)
Cochlear Implantation , Cochlear Implants , Deafness , Persons With Hearing Impairments , Speech Perception , Child , Humans , Speech Intelligibility , Deafness/surgery , Reproducibility of Results
SELECTION OF CITATIONS
SEARCH DETAIL
...