Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 32.857
Filter
1.
Nat Commun ; 15(1): 4835, 2024 Jun 06.
Article in English | MEDLINE | ID: mdl-38844457

ABSTRACT

Humans produce two forms of cognitively complex vocalizations: speech and song. It is debated whether these differ based primarily on culturally specific, learned features, or if acoustical features can reliably distinguish them. We study the spectro-temporal modulation patterns of vocalizations produced by 369 people living in 21 urban, rural, and small-scale societies across six continents. Specific ranges of spectral and temporal modulations, overlapping within categories and across societies, significantly differentiate speech from song. Machine-learning classification shows that this effect is cross-culturally robust, vocalizations being reliably classified solely from their spectro-temporal features across all 21 societies. Listeners unfamiliar with the cultures classify these vocalizations using similar spectro-temporal cues as the machine learning algorithm. Finally, spectro-temporal features are better able to discriminate song from speech than a broad range of other acoustical variables, suggesting that spectro-temporal modulation-a key feature of auditory neuronal tuning-accounts for a fundamental difference between these categories.


Subject(s)
Machine Learning , Speech , Humans , Speech/physiology , Male , Female , Adult , Acoustics , Cross-Cultural Comparison , Auditory Perception/physiology , Sound Spectrography , Singing/physiology , Music , Middle Aged , Young Adult
2.
Proc Natl Acad Sci U S A ; 121(24): e2311570121, 2024 Jun 11.
Article in English | MEDLINE | ID: mdl-38830095

ABSTRACT

Even a transient period of hearing loss during the developmental critical period can induce long-lasting deficits in temporal and spectral perception. These perceptual deficits correlate with speech perception in humans. In gerbils, these hearing loss-induced perceptual deficits are correlated with a reduction of both ionotropic GABAA and metabotropic GABAB receptor-mediated synaptic inhibition in auditory cortex, but most research on critical period plasticity has focused on GABAA receptors. Therefore, we developed viral vectors to express proteins that would upregulate gerbil postsynaptic inhibitory receptor subunits (GABAA, Gabra1; GABAB, Gabbr1b) in pyramidal neurons, and an enzyme that mediates GABA synthesis (GAD65) presynaptically in parvalbumin-expressing interneurons. A transient period of developmental hearing loss during the auditory critical period significantly impaired perceptual performance on two auditory tasks: amplitude modulation depth detection and spectral modulation depth detection. We then tested the capacity of each vector to restore perceptual performance on these auditory tasks. While both GABA receptor vectors increased the amplitude of cortical inhibitory postsynaptic potentials, only viral expression of postsynaptic GABAB receptors improved perceptual thresholds to control levels. Similarly, presynaptic GAD65 expression improved perceptual performance on spectral modulation detection. These findings suggest that recovering performance on auditory perceptual tasks depends on GABAB receptor-dependent transmission at the auditory cortex parvalbumin to pyramidal synapse and point to potential therapeutic targets for developmental sensory disorders.


Subject(s)
Auditory Cortex , Gerbillinae , Hearing Loss , Animals , Auditory Cortex/metabolism , Auditory Cortex/physiopathology , Hearing Loss/genetics , Hearing Loss/physiopathology , Receptors, GABA-B/metabolism , Receptors, GABA-B/genetics , Glutamate Decarboxylase/metabolism , Glutamate Decarboxylase/genetics , Receptors, GABA-A/metabolism , Receptors, GABA-A/genetics , Parvalbumins/metabolism , Parvalbumins/genetics , Auditory Perception/physiology , Pyramidal Cells/metabolism , Pyramidal Cells/physiology , Genetic Vectors/genetics
3.
J Acoust Soc Am ; 155(6): 3639-3653, 2024 Jun 01.
Article in English | MEDLINE | ID: mdl-38836771

ABSTRACT

The estimation of auditory evoked potentials requires deconvolution when the duration of the responses to be recovered exceeds the inter-stimulus interval. Based on least squares deconvolution, in this article we extend the procedure to the case of a multi-response convolutional model, that is, a model in which different categories of stimulus are expected to evoke different responses. The computational cost of the multi-response deconvolution significantly increases with the number of responses to be deconvolved, which restricts its applicability in practical situations. In order to alleviate this restriction, we propose to perform the multi-response deconvolution in a reduced representation space associated with a latency-dependent filtering of auditory responses, which provides a significant dimensionality reduction. We demonstrate the practical viability of the multi-response deconvolution with auditory responses evoked by clicks presented at different levels and categorized according to their stimulation level. The multi-response deconvolution applied in a reduced representation space provides the least squares estimation of the responses with a reasonable computational load. matlab/Octave code implementing the proposed procedure is included as supplementary material.


Subject(s)
Acoustic Stimulation , Evoked Potentials, Auditory , Evoked Potentials, Auditory/physiology , Humans , Acoustic Stimulation/methods , Male , Adult , Electroencephalography/methods , Female , Least-Squares Analysis , Young Adult , Signal Processing, Computer-Assisted , Reaction Time , Auditory Perception/physiology
4.
J Neurodev Disord ; 16(1): 28, 2024 Jun 03.
Article in English | MEDLINE | ID: mdl-38831410

ABSTRACT

BACKGROUND: In the search for objective tools to quantify neural function in Rett Syndrome (RTT), which are crucial in the evaluation of therapeutic efficacy in clinical trials, recordings of sensory-perceptual functioning using event-related potential (ERP) approaches have emerged as potentially powerful tools. Considerable work points to highly anomalous auditory evoked potentials (AEPs) in RTT. However, an assumption of the typical signal-averaging method used to derive these measures is "stationarity" of the underlying responses - i.e. neural responses to each input are highly stereotyped. An alternate possibility is that responses to repeated stimuli are highly variable in RTT. If so, this will significantly impact the validity of assumptions about underlying neural dysfunction, and likely lead to overestimation of underlying neuropathology. To assess this possibility, analyses at the single-trial level assessing signal-to-noise ratios (SNR), inter-trial variability (ITV) and inter-trial phase coherence (ITPC) are necessary. METHODS: AEPs were recorded to simple 100 Hz tones from 18 RTT and 27 age-matched controls (Ages: 6-22 years). We applied standard AEP averaging, as well as measures of neuronal reliability at the single-trial level (i.e. SNR, ITV, ITPC). To separate signal-carrying components from non-neural noise sources, we also applied a denoising source separation (DSS) algorithm and then repeated the reliability measures. RESULTS: Substantially increased ITV, lower SNRs, and reduced ITPC were observed in auditory responses of RTT participants, supporting a "neural unreliability" account. Application of the DSS technique made it clear that non-neural noise sources contribute to overestimation of the extent of processing deficits in RTT. Post-DSS, ITV measures were substantially reduced, so much so that pre-DSS ITV differences between RTT and TD populations were no longer detected. In the case of SNR and ITPC, DSS substantially improved these estimates in the RTT population, but robust differences between RTT and TD were still fully evident. CONCLUSIONS: To accurately represent the degree of neural dysfunction in RTT using the ERP technique, a consideration of response reliability at the single-trial level is highly advised. Non-neural sources of noise lead to overestimation of the degree of pathological processing in RTT, and denoising source separation techniques during signal processing substantially ameliorate this issue.


Subject(s)
Electroencephalography , Evoked Potentials, Auditory , Rett Syndrome , Humans , Rett Syndrome/physiopathology , Rett Syndrome/complications , Adolescent , Female , Evoked Potentials, Auditory/physiology , Child , Young Adult , Auditory Perception/physiology , Reproducibility of Results , Acoustic Stimulation , Male , Signal-To-Noise Ratio , Adult
5.
J Acoust Soc Am ; 155(6): 3715-3729, 2024 Jun 01.
Article in English | MEDLINE | ID: mdl-38847595

ABSTRACT

Emerging technologies of virtual reality (VR) and augmented reality (AR) are enhancing soundscape research, potentially producing new insights by enabling controlled conditions while preserving the context of a virtual gestalt within the soundscape concept. This study explored the ecological validity of virtual environments for subjective evaluations in soundscape research, focusing on the authenticity of virtual audio-visual environments for reproducibility. Different technologies for creating and reproducing virtual environments were compared, including field recording, simulated VR, AR, and audio-only presentation, in two audio-visual reproduction settings, a head-mounted display with head-tracked headphones and a VR lab with head-locked headphones. Via a series of soundwalk- and lab-based experiments, the results indicate that field recording technologies provided the most authentic audio-visual environments, followed by AR, simulated VR, and audio-only approaches. The authenticity level influenced subjective evaluations of virtual environments, e.g., arousal/eventfulness and pleasantness. The field recording and AR-based technologies closely matched the on-site soundwalk ratings in arousal, while the other approaches scored lower. All the approaches had significantly lower pleasantness ratings compared to on-site evaluations. The choice of audio-visual reproduction technology did not significantly impact the evaluations. Overall, the results suggest virtual environments with high authenticity can be useful for future soundscape research and design.


Subject(s)
Auditory Perception , Virtual Reality , Humans , Female , Male , Adult , Young Adult , Augmented Reality , Acoustic Stimulation , Sound , Reproducibility of Results
6.
J Acoust Soc Am ; 155(6): 3742-3759, 2024 Jun 01.
Article in English | MEDLINE | ID: mdl-38856312

ABSTRACT

Amplitude modulation (AM) of a masker reduces its masking on a simultaneously presented unmodulated pure-tone target, which likely involves dip listening. This study tested the idea that dip-listening efficiency may depend on stimulus context, i.e., the match in AM peakedness (AMP) between the masker and a precursor or postcursor stimulus, assuming a form of temporal pattern analysis process. Masked thresholds were measured in normal-hearing listeners using Schroeder-phase harmonic complexes as maskers and precursors or postcursors. Experiment 1 showed threshold elevation (i.e., interference) when a flat cursor preceded or followed a peaked masker, suggesting proactive and retroactive temporal pattern analysis. Threshold decline (facilitation) was observed when the masker AMP was matched to the precursor, irrespective of stimulus AMP, suggesting only proactive processing. Subsequent experiments showed that both interference and facilitation (1) remained robust when a temporal gap was inserted between masker and cursor, (2) disappeared when an F0-difference was introduced between masker and precursor, and (3) decreased when the presentation level was reduced. These results suggest an important role of envelope regularity in dip listening, especially when masker and cursor are F0-matched and, therefore, form one perceptual stream. The reported effects seem to represent a time-domain variant of comodulation masking release.


Subject(s)
Acoustic Stimulation , Auditory Threshold , Perceptual Masking , Humans , Young Adult , Adult , Time Factors , Female , Male , Audiometry, Pure-Tone , Auditory Perception/physiology
7.
Brain Behav ; 14(6): e3571, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38841736

ABSTRACT

OBJECTIVE: This study aims to control all hearing thresholds, including extended high frequencies (EHFs), presents stimuli of varying difficulty levels, and measures electroencephalography (EEG) and pupillometry responses to determine whether listening difficulty in tinnitus patients is effort or fatigue-related. METHODS: Twenty-one chronic tinnitus patients and 26 matched healthy controls having normal pure-tone averages with symmetrical hearing thresholds were included. Subjects were evaluated with 0.125-20 kHz pure-tone audiometry, Montreal Cognitive Assessment Test (MoCA), Tinnitus Handicap Inventory (THI), EEG, and pupillometry. RESULTS: Pupil dilatation and EEG alpha power during the "encoding" phase of the presented sentence in tinnitus patients were less in all listening conditions (p < .05). Also, there was no statistically significant relationship between EEG and pupillometry components for all listening conditions and THI or MoCA (p > .05). CONCLUSION: EEG and pupillometry results under various listening conditions indicate potential listening effort in tinnitus patients even if all frequencies, including EHFs, are controlled. Also, we suggest that pupillometry should be interpreted with caution in autonomic nervous system-related conditions such as tinnitus.


Subject(s)
Electroencephalography , Pupil , Tinnitus , Humans , Tinnitus/physiopathology , Tinnitus/diagnosis , Male , Female , Electroencephalography/methods , Adult , Middle Aged , Pupil/physiology , Audiometry, Pure-Tone , Auditory Perception/physiology , Auditory Threshold/physiology
8.
JASA Express Lett ; 4(5)2024 May 01.
Article in English | MEDLINE | ID: mdl-38717467

ABSTRACT

A long-standing quest in audition concerns understanding relations between behavioral measures and neural representations of changes in sound intensity. Here, we examined relations between aspects of intensity perception and central neural responses within the inferior colliculus of unanesthetized rabbits (by averaging the population's spike count/level functions). We found parallels between the population's neural output and: (1) how loudness grows with intensity; (2) how loudness grows with duration; (3) how discrimination of intensity improves with increasing sound level; (4) findings that intensity discrimination does not depend on duration; and (5) findings that duration discrimination is a constant fraction of base duration.


Subject(s)
Inferior Colliculi , Loudness Perception , Animals , Rabbits , Loudness Perception/physiology , Inferior Colliculi/physiology , Acoustic Stimulation/methods , Discrimination, Psychological/physiology , Auditory Perception/physiology , Neurons/physiology
9.
JASA Express Lett ; 4(5)2024 May 01.
Article in English | MEDLINE | ID: mdl-38727569

ABSTRACT

Bimodal stimulation, a cochlear implant (CI) in one ear and a hearing aid (HA) in the other, provides highly asymmetrical inputs. To understand how asymmetry affects perception and memory, forward and backward digit spans were measured in nine bimodal listeners. Spans were unchanged from monotic to diotic presentation; there was an average two-digit decrease for dichotic presentation with some extreme cases of decreases to zero spans. Interaurally asymmetrical decreases were not predicted based on the device or better-functioning ear. Therefore, bimodal listeners can demonstrate a strong ear dominance, diminishing memory recall dichotically even when perception was intact monaurally.


Subject(s)
Cochlear Implants , Humans , Middle Aged , Aged , Male , Female , Dichotic Listening Tests , Adult , Auditory Perception/physiology , Hearing Aids
10.
Brain Behav ; 14(5): e3517, 2024 May.
Article in English | MEDLINE | ID: mdl-38702896

ABSTRACT

INTRODUCTION: Attention and working memory are key cognitive functions that allow us to select and maintain information in our mind for a short time, being essential for our daily life and, in particular, for learning and academic performance. It has been shown that musical training can improve working memory performance, but it is still unclear if and how the neural mechanisms of working memory and particularly attention are implicated in this process. In this work, we aimed to identify the oscillatory signature of bimodal attention and working memory that contributes to improved working memory in musically trained children. MATERIALS AND METHODS: We recruited children with and without musical training and asked them to complete a bimodal (auditory/visual) attention and working memory task, whereas their brain activity was measured using electroencephalography. Behavioral, time-frequency, and source reconstruction analyses were made. RESULTS: Results showed that, overall, musically trained children performed better on the task than children without musical training. When comparing musically trained children with children without musical training, we found modulations in the alpha band pre-stimuli onset and the beginning of stimuli onset in the frontal and parietal regions. These correlated with correct responses to the attended modality. Moreover, during the end phase of stimuli presentation, we found modulations correlating with correct responses independent of attention condition in the theta and alpha bands, in the left frontal and right parietal regions. CONCLUSIONS: These results suggest that musically trained children have improved neuronal mechanisms for both attention allocation and memory encoding. Our results can be important for developing interventions for people with attention and working memory difficulties.


Subject(s)
Alpha Rhythm , Attention , Memory, Short-Term , Music , Theta Rhythm , Humans , Memory, Short-Term/physiology , Attention/physiology , Male , Female , Child , Theta Rhythm/physiology , Alpha Rhythm/physiology , Auditory Perception/physiology , Electroencephalography , Visual Perception/physiology , Brain/physiology
11.
J Acoust Soc Am ; 155(5): 3101-3117, 2024 May 01.
Article in English | MEDLINE | ID: mdl-38722101

ABSTRACT

Cochlear implant (CI) users often report being unsatisfied by music listening through their hearing device. Vibrotactile stimulation could help alleviate those challenges. Previous research has shown that musical stimuli was given higher preference ratings by normal-hearing listeners when concurrent vibrotactile stimulation was congruent in intensity and timing with the corresponding auditory signal compared to incongruent. However, it is not known whether this is also the case for CI users. Therefore, in this experiment, we presented 18 CI users and 24 normal-hearing listeners with five melodies and five different audio-to-tactile maps. Each map varied the congruence between the audio and tactile signals related to intensity, fundamental frequency, and timing. Participants were asked to rate the maps from zero to 100, based on preference. It was shown that almost all normal-hearing listeners, as well as a subset of the CI users, preferred tactile stimulation, which was congruent with the audio in intensity and timing. However, many CI users had no difference in preference between timing aligned and timing unaligned stimuli. The results provide evidence that vibrotactile music enjoyment enhancement could be a solution for some CI users; however, more research is needed to understand which CI users can benefit from it most.


Subject(s)
Acoustic Stimulation , Auditory Perception , Cochlear Implants , Music , Humans , Female , Male , Adult , Middle Aged , Aged , Auditory Perception/physiology , Young Adult , Patient Preference , Cochlear Implantation/instrumentation , Touch Perception/physiology , Vibration , Touch
12.
Brain Cogn ; 177: 106161, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38696928

ABSTRACT

Narrative comprehension relies on basic sensory processing abilities, such as visual and auditory processing, with recent evidence for utilizing executive functions (EF), which are also engaged during reading. EF was previously related to the "supporter" of engaging the auditory and visual modalities in different cognitive tasks, with evidence of lower efficiency in this process among those with reading difficulties in the absence of a visual stimulus (i.e. while listening to stories). The current study aims to fill out the gap related to the level of reliance on these neural circuits while visual aids (pictures) are involved during story listening in relation to reading skills. Functional MRI data were collected from 44 Hebrew-speaking children aged 8-12 years while listening to stories with vs without visual stimuli (i.e., pictures). Functional connectivity of networks supporting reading was defined in each condition and compared between the conditions against behavioral reading measures. Lower reading skills were related to greater functional connectivity values between EF networks (default mode and memory networks), and between the auditory and memory networks for the stories with vs without the visual stimulation. A greater difference in functional connectivity between the conditions was related to lower reading scores. We conclude that lower reading skills in children may be related to a need for greater scaffolding, i.e., visual stimulation such as pictures describing the narratives when listening to stories, which may guide future intervention approaches.


Subject(s)
Executive Function , Magnetic Resonance Imaging , Reading , Visual Perception , Humans , Child , Male , Female , Executive Function/physiology , Visual Perception/physiology , Auditory Perception/physiology , Comprehension/physiology , Photic Stimulation/methods , Nerve Net/physiology , Nerve Net/diagnostic imaging , Brain/physiology
13.
Trends Hear ; 28: 23312165241239541, 2024.
Article in English | MEDLINE | ID: mdl-38738337

ABSTRACT

Cochlear synaptopathy, a form of cochlear deafferentation, has been demonstrated in a number of animal species, including non-human primates. Both age and noise exposure contribute to synaptopathy in animal models, indicating that it may be a common type of auditory dysfunction in humans. Temporal bone and auditory physiological data suggest that age and occupational/military noise exposure also lead to synaptopathy in humans. The predicted perceptual consequences of synaptopathy include tinnitus, hyperacusis, and difficulty with speech-in-noise perception. However, confirming the perceptual impacts of this form of cochlear deafferentation presents a particular challenge because synaptopathy can only be confirmed through post-mortem temporal bone analysis and auditory perception is difficult to evaluate in animals. Animal data suggest that deafferentation leads to increased central gain, signs of tinnitus and abnormal loudness perception, and deficits in temporal processing and signal-in-noise detection. If equivalent changes occur in humans following deafferentation, this would be expected to increase the likelihood of developing tinnitus, hyperacusis, and difficulty with speech-in-noise perception. Physiological data from humans is consistent with the hypothesis that deafferentation is associated with increased central gain and a greater likelihood of tinnitus perception, while human data on the relationship between deafferentation and hyperacusis is extremely limited. Many human studies have investigated the relationship between physiological correlates of deafferentation and difficulty with speech-in-noise perception, with mixed findings. A non-linear relationship between deafferentation and speech perception may have contributed to the mixed results. When differences in sample characteristics and study measurements are considered, the findings may be more consistent.


Subject(s)
Cochlea , Speech Perception , Tinnitus , Humans , Cochlea/physiopathology , Tinnitus/physiopathology , Tinnitus/diagnosis , Animals , Speech Perception/physiology , Hyperacusis/physiopathology , Noise/adverse effects , Auditory Perception/physiology , Synapses/physiology , Hearing Loss, Noise-Induced/physiopathology , Hearing Loss, Noise-Induced/diagnosis , Loudness Perception
14.
Headache ; 64(5): 482-493, 2024 May.
Article in English | MEDLINE | ID: mdl-38693749

ABSTRACT

OBJECTIVE: In this cross-sectional observational study, we aimed to investigate sensory profiles and multisensory integration processes in women with migraine using virtual dynamic interaction systems. BACKGROUND: Compared to studies on unimodal sensory processing, fewer studies show that multisensory integration differs in patients with migraine. Multisensory integration of visual, auditory, verbal, and haptic modalities has not been evaluated in migraine. METHODS: A 12-min virtual dynamic interaction game consisting of four parts was played by the participants. During the game, the participants were exposed to either visual stimuli only or multisensory stimuli in which auditory, verbal, and haptic stimuli were added to the visual stimuli. A total of 78 women participants (28 with migraine without aura and 50 healthy controls) were enrolled in this prospective exploratory study. Patients with migraine and healthy participants who met the inclusion criteria were randomized separately into visual and multisensory groups: Migraine multisensory (14 adults), migraine visual (14 adults), healthy multisensory (25 adults), and healthy visual (25 adults). The Sensory Profile Questionnaire was utilized to assess the participants' sensory profiles. The game scores and survey results were analyzed. RESULTS: In visual stimulus, the gaming performance scores of patients with migraine without aura were similar to the healthy controls, at a median (interquartile range [IQR]) of 81.8 (79.5-85.8) and 80.9 (77.1-84.2) (p = 0.149). Error rate of visual stimulus in patients with migraine without aura were comparable to healthy controls, at a median (IQR) of 0.11 (0.08-0.13) and 0.12 (0.10-0.14), respectively (p = 0,166). In multisensory stimulation, average gaming score was lower in patients with migraine without aura compared to healthy individuals (median [IQR] 82.2 [78.8-86.3] vs. 78.6 [74.0-82.4], p = 0.028). In women with migraine, exposure to new sensory modality upon visual stimuli in the fourth, seventh, and tenth rounds (median [IQR] 78.1 [74.1-82.0], 79.7 [77.2-82.5], 76.5 [70.2-82.1]) exhibited lower game scores compared to visual stimuli only (median [IQR] 82.3 [77.9-87.8], 84.2 [79.7-85.6], 80.8 [79.0-85.7], p = 0.044, p = 0.049, p = 0.016). According to the Sensory Profile Questionnaire results, sensory sensitivity, and sensory avoidance scores of patients with migraine (median [IQR] score 45.5 [41.0-54.7] and 47.0 [41.5-51.7]) were significantly higher than healthy participants (median [IQR] score 39.0 [34.0-44.2] and 40.0 [34.0-48.0], p < 0.001, p = 0.001). CONCLUSION: The virtual dynamic game approach showed for the first time that the gaming performance of patients with migraine without aura was negatively affected by the addition of auditory, verbal, and haptic stimuli onto visual stimuli. Multisensory integration of sensory modalities including haptic stimuli is disturbed even in the interictal period in women with migraine. Virtual games can be employed to assess the impact of sensory problems in the course of the disease. Also, sensory training could be a potential therapy target to improve multisensory processing in migraine.


Subject(s)
Migraine Disorders , Humans , Female , Adult , Cross-Sectional Studies , Migraine Disorders/physiopathology , Prospective Studies , Video Games , Visual Perception/physiology , Young Adult , Virtual Reality , Photic Stimulation/methods , Auditory Perception/physiology
15.
Philos Trans R Soc Lond B Biol Sci ; 379(1905): 20230186, 2024 Jul 08.
Article in English | MEDLINE | ID: mdl-38768210

ABSTRACT

Communication takes place within a network of multiple signallers and receivers. Social network analysis provides tools to quantify how an individual's social positioning affects group dynamics and the subsequent biological consequences. However, network analysis is rarely applied to animal communication, likely due to the logistical difficulties of monitoring natural communication networks. We generated a simulated communication network to investigate how variation in individual communication behaviours generates network effects, and how this communication network's structure feeds back to affect future signalling interactions. We simulated competitive acoustic signalling interactions among chorusing individuals and varied several parameters related to communication and chorus size to examine their effects on calling output and social connections. Larger choruses had higher noise levels, and this reduced network density and altered the relationships between individual traits and communication network position. Hearing sensitivity interacted with chorus size to affect both individuals' positions in the network and the acoustic output of the chorus. Physical proximity to competitors influenced signalling, but a distinctive communication network structure emerged when signal active space was limited. Our model raises novel predictions about communication networks that could be tested experimentally and identifies aspects of information processing in complex environments that remain to be investigated. This article is part of the theme issue 'The power of sound: unravelling how acoustic communication shapes group dynamics'.


Subject(s)
Auditory Perception , Animals , Auditory Perception/physiology , Vocalization, Animal/physiology , Animal Communication , Models, Biological , Birds/physiology , Acoustics , Social Behavior
16.
eNeuro ; 11(5)2024 May.
Article in English | MEDLINE | ID: mdl-38702187

ABSTRACT

Mismatch negativity (MMN) is commonly recognized as a neural signal of prediction error evoked by deviants from the expected patterns of sensory input. Studies show that MMN diminishes when sequence patterns become more predictable over a longer timescale. This implies that MMN is composed of multiple subcomponents, each responding to different levels of temporal regularities. To probe the hypothesized subcomponents in MMN, we record human electroencephalography during an auditory local-global oddball paradigm where the tone-to-tone transition probability (local regularity) and the overall sequence probability (global regularity) are manipulated to control temporal predictabilities at two hierarchical levels. We find that the size of MMN is correlated with both probabilities and the spatiotemporal structure of MMN can be decomposed into two distinct subcomponents. Both subcomponents appear as negative waveforms, with one peaking early in the central-frontal area and the other late in a more frontal area. With a quantitative predictive coding model, we map the early and late subcomponents to the prediction errors that are tied to local and global regularities, respectively. Our study highlights the hierarchical complexity of MMN and offers an experimental and analytical platform for developing a multitiered neural marker applicable in clinical settings.


Subject(s)
Acoustic Stimulation , Electroencephalography , Evoked Potentials, Auditory , Humans , Male , Female , Electroencephalography/methods , Young Adult , Adult , Evoked Potentials, Auditory/physiology , Acoustic Stimulation/methods , Auditory Perception/physiology , Brain/physiology , Brain Mapping , Adolescent
17.
J Neural Eng ; 21(3)2024 May 22.
Article in English | MEDLINE | ID: mdl-38729132

ABSTRACT

Objective.This study develops a deep learning (DL) method for fast auditory attention decoding (AAD) using electroencephalography (EEG) from listeners with hearing impairment (HI). It addresses three classification tasks: differentiating noise from speech-in-noise, classifying the direction of attended speech (left vs. right) and identifying the activation status of hearing aid noise reduction algorithms (OFF vs. ON). These tasks contribute to our understanding of how hearing technology influences auditory processing in the hearing-impaired population.Approach.Deep convolutional neural network (DCNN) models were designed for each task. Two training strategies were employed to clarify the impact of data splitting on AAD tasks: inter-trial, where the testing set used classification windows from trials that the training set had not seen, and intra-trial, where the testing set used unseen classification windows from trials where other segments were seen during training. The models were evaluated on EEG data from 31 participants with HI, listening to competing talkers amidst background noise.Main results.Using 1 s classification windows, DCNN models achieve accuracy (ACC) of 69.8%, 73.3% and 82.9% and area-under-curve (AUC) of 77.2%, 80.6% and 92.1% for the three tasks respectively on inter-trial strategy. In the intra-trial strategy, they achieved ACC of 87.9%, 80.1% and 97.5%, along with AUC of 94.6%, 89.1%, and 99.8%. Our DCNN models show good performance on short 1 s EEG samples, making them suitable for real-world applications. Conclusion: Our DCNN models successfully addressed three tasks with short 1 s EEG windows from participants with HI, showcasing their potential. While the inter-trial strategy demonstrated promise for assessing AAD, the intra-trial approach yielded inflated results, underscoring the important role of proper data splitting in EEG-based AAD tasks.Significance.Our findings showcase the promising potential of EEG-based tools for assessing auditory attention in clinical contexts and advancing hearing technology, while also promoting further exploration of alternative DL architectures and their potential constraints.


Subject(s)
Attention , Auditory Perception , Deep Learning , Electroencephalography , Hearing Loss , Humans , Attention/physiology , Female , Electroencephalography/methods , Male , Middle Aged , Hearing Loss/physiopathology , Hearing Loss/rehabilitation , Hearing Loss/diagnosis , Aged , Auditory Perception/physiology , Noise , Adult , Hearing Aids , Speech Perception/physiology , Neural Networks, Computer
18.
Multisens Res ; 37(2): 89-124, 2024 Feb 13.
Article in English | MEDLINE | ID: mdl-38714311

ABSTRACT

Prior studies investigating the effects of routine action video game play have demonstrated improvements in a variety of cognitive processes, including improvements in attentional tasks. However, there is little evidence indicating that the cognitive benefits of playing action video games generalize from simplified unisensory stimuli to multisensory scenes - a fundamental characteristic of natural, everyday life environments. The present study addressed if video game experience has an impact on crossmodal congruency effects when searching through such multisensory scenes. We compared the performance of action video game players (AVGPs) and non-video game players (NVGPs) on a visual search task for objects embedded in video clips of realistic scenes. We conducted two identical online experiments with gender-balanced samples, for a total of N = 130. Overall, the data replicated previous findings reporting search benefits when visual targets were accompanied by semantically congruent auditory events, compared to neutral or incongruent ones. However, according to the results, AVGPs did not consistently outperform NVGPs in the overall search task, nor did they use multisensory cues more efficiently than NVGPs. Exploratory analyses with self-reported gender as a variable revealed a potential difference in response strategy between experienced male and female AVGPs when dealing with crossmodal cues. These findings suggest that the generalization of the advantage of AVG experience to realistic, crossmodal situations should be made with caution and considering gender-related issues.


Subject(s)
Attention , Video Games , Visual Perception , Humans , Male , Female , Visual Perception/physiology , Young Adult , Adult , Attention/physiology , Auditory Perception/physiology , Photic Stimulation , Adolescent , Reaction Time/physiology , Cues , Acoustic Stimulation
19.
Multisens Res ; 37(2): 143-162, 2024 Apr 30.
Article in English | MEDLINE | ID: mdl-38714315

ABSTRACT

A vital heuristic used when making judgements on whether audio-visual signals arise from the same event, is the temporal coincidence of the respective signals. Previous research has highlighted a process, whereby the perception of simultaneity rapidly recalibrates to account for differences in the physical temporal offsets of stimuli. The current paper investigated whether rapid recalibration also occurs in response to differences in central arrival latencies, driven by visual-intensity-dependent processing times. In a behavioural experiment, observers completed a temporal-order judgement (TOJ), simultaneity judgement (SJ) and simple reaction-time (RT) task and responded to audio-visual trials that were preceded by other audio-visual trials with either a bright or dim visual stimulus. It was found that the point of subjective simultaneity shifted, due to the visual intensity of the preceding stimulus, in the TOJ, but not SJ task, while the RT data revealed no effect of preceding intensity. Our data therefore provide some evidence that the perception of simultaneity rapidly recalibrates based on stimulus intensity.


Subject(s)
Acoustic Stimulation , Auditory Perception , Photic Stimulation , Reaction Time , Visual Perception , Humans , Visual Perception/physiology , Auditory Perception/physiology , Male , Female , Reaction Time/physiology , Adult , Young Adult , Judgment/physiology
20.
Codas ; 36(2): e20230048, 2024.
Article in Portuguese, English | MEDLINE | ID: mdl-38695432

ABSTRACT

PURPOSE: To correlate behavioral assessment results of central auditory processing and the self-perception questionnaire after acoustically controlled auditory training. METHODS: The study assessed 10 individuals with a mean age of 44.5 years who had suffered mild traumatic brain injury. They underwent behavioral assessment of central auditory processing and answered the Formal Auditory Training self-perception questionnaire after the therapeutic intervention - whose questions address auditory perception, understanding orders, request to repeat statements, occurrence of misunderstandings, attention span, auditory performance in noisy environments, telephone communication, and self-esteem. Patients were asked to indicate the frequency with which the listed behaviors occurred. RESULTS: Figure-ground, sequential memory for sounds, and temporal processing correlated with improvement in following instructions, fewer requests to repeat statements, increased attention span, improved communication, and understanding on the phone and when watching TV. CONCLUSION: Auditory closure, figure-ground, and temporal processing had improved in the assessment after the acoustically controlled auditory training, and there were fewer auditory behavior complaints.


OBJETIVO: Correlacionar os resultados da avaliação comportamental do processamento auditivo central e do questionário de autopercepção após o treinamento auditivo acusticamente controlado. MÉTODO: Foram avaliados dez indivíduos com média de idade de 44,5 anos, que sofreram traumatismo cranioencefálico de grau leve. Os indivíduos foram submetidos a avaliação comportamental do processamento auditivo central e também responderam ao questionário de autopercepção "Treinamento Auditivo Formal" após a intervenção terapêutica. O questionário foi composto por questões referentes a percepção auditiva, compreensão de ordens, solicitação de repetição de enunciados, ocorrência mal-entendidos, tempo de atenção, desempenho auditivo em ambiente ruidoso, comunicação ao telefone e autoestima e os pacientes foram solicitados a assinalar a frequência de ocorrência dos comportamentos listados. RESULTADOS: As habilidades auditivas de figura-fundo e memória para sons em sequência e processamento temporal correlacionaram-se com melhora para seguir instruções, diminuição das solicitações de repetições e aumento do tempo de atenção e melhora da comunicação e da compreensão ao telefone e para assistir TV. CONCLUSÃO: Observou-se adequação das habilidades auditivas de fechamento auditivo, figura fundo, e processamento temporal na avaliação pós-treinamento auditivo acusticamente controlado, além de redução das queixas quanto ao comportamento auditivo.


Subject(s)
Auditory Perception , Self Concept , Humans , Adult , Male , Female , Auditory Perception/physiology , Surveys and Questionnaires , Middle Aged , Brain Concussion/psychology , Brain Concussion/rehabilitation , Acoustic Stimulation/methods , Young Adult
SELECTION OF CITATIONS
SEARCH DETAIL
...