Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 90
Filter
Add more filters










Publication year range
1.
PLoS One ; 19(4): e0301098, 2024.
Article in English | MEDLINE | ID: mdl-38557696

ABSTRACT

We present Audiovisual Moments in Time (AVMIT), a large-scale dataset of audiovisual action events. In an extensive annotation task 11 participants labelled a subset of 3-second audiovisual videos from the Moments in Time dataset (MIT). For each trial, participants assessed whether the labelled audiovisual action event was present and whether it was the most prominent feature of the video. The dataset includes the annotation of 57,177 audiovisual videos, each independently evaluated by 3 of 11 trained participants. From this initial collection, we created a curated test set of 16 distinct action classes, with 60 videos each (960 videos). We also offer 2 sets of pre-computed audiovisual feature embeddings, using VGGish/YamNet for audio data and VGG16/EfficientNetB0 for visual data, thereby lowering the barrier to entry for audiovisual DNN research. We explored the advantages of AVMIT annotations and feature embeddings to improve performance on audiovisual event recognition. A series of 6 Recurrent Neural Networks (RNNs) were trained on either AVMIT-filtered audiovisual events or modality-agnostic events from MIT, and then tested on our audiovisual test set. In all RNNs, top 1 accuracy was increased by 2.71-5.94% by training exclusively on audiovisual events, even outweighing a three-fold increase in training data. Additionally, we introduce the Supervised Audiovisual Correspondence (SAVC) task whereby a classifier must discern whether audio and visual streams correspond to the same action label. We trained 6 RNNs on the SAVC task, with or without AVMIT-filtering, to explore whether AVMIT is helpful for cross-modal learning. In all RNNs, accuracy improved by 2.09-19.16% with AVMIT-filtered data. We anticipate that the newly annotated AVMIT dataset will serve as a valuable resource for research and comparative experiments involving computational models and human participants, specifically when addressing research questions where audiovisual correspondence is of critical importance.


Subject(s)
Learning , Neural Networks, Computer , Humans
2.
Hum Brain Mapp ; 45(4): e26653, 2024 Mar.
Article in English | MEDLINE | ID: mdl-38488460

ABSTRACT

Face-to-face communication relies on the integration of acoustic speech signals with the corresponding facial articulations. In the McGurk illusion, an auditory /ba/ phoneme presented simultaneously with a facial articulation of a /ga/ (i.e., viseme), is typically fused into an illusory 'da' percept. Despite its widespread use as an index of audiovisual speech integration, critics argue that it arises from perceptual processes that differ categorically from natural speech recognition. Conversely, Bayesian theoretical frameworks suggest that both the illusory McGurk and the veridical audiovisual congruent speech percepts result from probabilistic inference based on noisy sensory signals. According to these models, the inter-sensory conflict in McGurk stimuli may only increase observers' perceptual uncertainty. This functional magnetic resonance imaging (fMRI) study presented participants (20 male and 24 female) with audiovisual congruent, McGurk (i.e., auditory /ba/ + visual /ga/), and incongruent (i.e., auditory /ga/ + visual /ba/) stimuli along with their unisensory counterparts in a syllable categorization task. Behaviorally, observers' response entropy was greater for McGurk compared to congruent audiovisual stimuli. At the neural level, McGurk stimuli increased activations in a widespread neural system, extending from the inferior frontal sulci (IFS) to the pre-supplementary motor area (pre-SMA) and insulae, typically involved in cognitive control processes. Crucially, in line with Bayesian theories these activation increases were fully accounted for by observers' perceptual uncertainty as measured by their response entropy. Our findings suggest that McGurk and congruent speech processing rely on shared neural mechanisms, thereby supporting the McGurk illusion as a valid measure of natural audiovisual speech perception.


Subject(s)
Illusions , Speech Perception , Humans , Male , Female , Auditory Perception/physiology , Speech/physiology , Illusions/physiology , Visual Perception/physiology , Bayes Theorem , Uncertainty , Speech Perception/physiology , Acoustic Stimulation/methods , Photic Stimulation/methods
3.
PLoS Biol ; 22(2): e3002494, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38319934

ABSTRACT

Effective interactions with the environment rely on the integration of multisensory signals: Our brains must efficiently combine signals that share a common source, and segregate those that do not. Healthy ageing can change or impair this process. This functional magnetic resonance imaging study assessed the neural mechanisms underlying age differences in the integration of auditory and visual spatial cues. Participants were presented with synchronous audiovisual signals at various degrees of spatial disparity and indicated their perceived sound location. Behaviourally, older adults were able to maintain localisation accuracy. At the neural level, they integrated auditory and visual cues into spatial representations along dorsal auditory and visual processing pathways similarly to their younger counterparts but showed greater activations in a widespread system of frontal, temporal, and parietal areas. According to multivariate Bayesian decoding, these areas encoded critical stimulus information beyond that which was encoded in the brain areas commonly activated by both groups. Surprisingly, however, the boost in information provided by these areas with age-related activation increases was comparable across the 2 age groups. This dissociation-between comparable information encoded in brain activation patterns across the 2 age groups, but age-related increases in regional blood-oxygen-level-dependent responses-contradicts the widespread notion that older adults recruit new regions as a compensatory mechanism to encode task-relevant information. Instead, our findings suggest that activation increases in older adults reflect nonspecific or modulatory mechanisms related to less efficient or slower processing, or greater demands on attentional resources.


Subject(s)
Brain Mapping , Visual Perception , Humans , Aged , Bayes Theorem , Visual Perception/physiology , Brain/physiology , Attention/physiology , Acoustic Stimulation/methods , Auditory Perception/physiology , Photic Stimulation/methods , Magnetic Resonance Imaging
4.
J Cogn Neurosci ; 36(4): 730-733, 2024 04 01.
Article in English | MEDLINE | ID: mdl-38307128

ABSTRACT

The papers collected in this Special Focus, prompted by S. Buergers and U. Noppeney [The role of alpha oscillations in temporal binding within and across the senses. Nature Human Behaviour, 6, 732-742, 2022], have raised several interesting ideas, arguments, and empirical results relating to the alpha temporal resolution hypothesis. Here we briefly respond to these, and in the process emphasize four challenges for future research: defining the scope and limitation of the hypothesis; developing experimental paradigms and study designs that rigorously test its tenets; decomposing the scalp-level signal and isolating underlying neural circuits; and bringing uniformity to the current diversity of analysis and statistical methods. Addressing these challenges will facilitate the progression from merely correlating alpha frequency with various perceptual phenomena to establishing whether and (if so) how alpha frequency influences sensory integration and segregation.


Subject(s)
Visual Perception , Humans , Visual Perception/physiology
5.
J Cogn Neurosci ; 36(4): 655-690, 2024 04 01.
Article in English | MEDLINE | ID: mdl-38330177

ABSTRACT

An intriguing question in cognitive neuroscience is whether alpha oscillations shape how the brain transforms the continuous sensory inputs into distinct percepts. According to the alpha temporal resolution hypothesis, sensory signals arriving within a single alpha cycle are integrated, whereas those in separate cycles are segregated. Consequently, shorter alpha cycles should be associated with smaller temporal binding windows and higher temporal resolution. However, the evidence supporting this hypothesis is contentious, and the neural mechanisms remain unclear. In this review, we first elucidate the alpha temporal resolution hypothesis and the neural circuitries that generate alpha oscillations. We then critically evaluate study designs, experimental paradigms, psychophysics, and neurophysiological analyses that have been employed to investigate the role of alpha frequency in temporal binding. Through the lens of this methodological framework, we then review evidence from between-subject, within-subject, and causal perturbation studies. Our review highlights the inherent interpretational ambiguities posed by previous study designs and experimental paradigms and the extensive variability in analysis choices across studies. We also suggest best practice recommendations that may help to guide future research. To establish a mechanistic role of alpha frequency in temporal parsing, future research is needed that demonstrates its causal effects on the temporal binding window with consistent, experimenter-independent methods.


Subject(s)
Brain , Visual Perception , Humans , Visual Perception/physiology , Brain/physiology , Research Design
6.
Adv Exp Med Biol ; 1437: 59-76, 2024.
Article in English | MEDLINE | ID: mdl-38270853

ABSTRACT

Multisensory perception is critical for effective interaction with the environment, but human responses to multisensory stimuli vary across the lifespan and appear changed in some atypical populations. In this review chapter, we consider multisensory integration within a normative Bayesian framework. We begin by outlining the complex computational challenges of multisensory causal inference and reliability-weighted cue integration, and discuss whether healthy young adults behave in accordance with normative Bayesian models. We then compare their behaviour with various other human populations (children, older adults, and those with neurological or neuropsychiatric disorders). In particular, we consider whether the differences seen in these groups are due only to changes in their computational parameters (such as sensory noise or perceptual priors), or whether the fundamental computational principles (such as reliability weighting) underlying multisensory perception may also be altered. We conclude by arguing that future research should aim explicitly to differentiate between these possibilities.


Subject(s)
Health Status , Longevity , Child , Young Adult , Humans , Aged , Bayes Theorem , Reproducibility of Results , Causality
7.
Philos Trans R Soc Lond B Biol Sci ; 378(1886): 20220332, 2023 09 25.
Article in English | MEDLINE | ID: mdl-37545306

ABSTRACT

Sensory systems evolved to provide the organism with information about the environment to guide adaptive behaviour. Neuroscientists and psychologists have traditionally considered each sense independently, a legacy of Aristotle and a natural consequence of their distinct physical and anatomical bases. However, from the point of view of the organism, perception and sensorimotor behaviour are fundamentally multi-modal; after all, each modality provides complementary information about the same world. Classic studies revealed much about where and how sensory signals are combined to improve performance, but these tended to treat multisensory integration as a static, passive, bottom-up process. It has become increasingly clear how this approach falls short, ignoring the interplay between perception and action, the temporal dynamics of the decision process and the many ways by which the brain can exert top-down control of integration. The goal of this issue is to highlight recent advances on these higher order aspects of multisensory processing, which together constitute a mainstay of our understanding of complex, natural behaviour and its neural basis. This article is part of the theme issue 'Decision and control processes in multisensory perception'.


Subject(s)
Brain , Decision Making , Visual Perception
8.
Philos Trans R Soc Lond B Biol Sci ; 378(1886): 20220348, 2023 09 25.
Article in English | MEDLINE | ID: mdl-37545307

ABSTRACT

Almost all decisions in everyday life rely on multiple sensory inputs that can come from common or independent causes. These situations invoke perceptual uncertainty about environmental properties and the signals' causal structure. Using the audiovisual McGurk illusion, this study investigated how observers formed perceptual and causal confidence judgements in information integration tasks under causal uncertainty. Observers were presented with spoken syllables, their corresponding articulatory lip movements or their congruent and McGurk combinations (e.g. auditory B/P with visual G/K). Observers reported their perceived auditory syllable, the causal structure and confidence for each judgement. Observers were more accurate and confident on congruent than unisensory trials. Their perceptual and causal confidence were tightly related over trials as predicted by the interactive nature of perceptual and causal inference. Further, observers assigned comparable perceptual and causal confidence to veridical 'G/K' percepts on audiovisual congruent trials and their causal and perceptual metamers on McGurk trials (i.e. illusory 'G/K' percepts). Thus, observers metacognitively evaluate the integrated audiovisual percept with limited access to the conflicting unisensory stimulus components on McGurk trials. Collectively, our results suggest that observers form meaningful perceptual and causal confidence judgements about multisensory scenes that are qualitatively consistent with principles of Bayesian causal inference. This article is part of the theme issue 'Decision and control processes in multisensory perception'.


Subject(s)
Illusions , Metacognition , Speech Perception , Humans , Auditory Perception , Visual Perception , Bayes Theorem , Photic Stimulation , Acoustic Stimulation
9.
Nat Commun ; 13(1): 3924, 2022 07 07.
Article in English | MEDLINE | ID: mdl-35798733

ABSTRACT

The brain adapts dynamically to the changing sensory statistics of its environment. Recent research has started to delineate the neural circuitries and representations that support this cross-sensory plasticity. Combining psychophysics and model-based representational fMRI and EEG we characterized how the adult human brain adapts to misaligned audiovisual signals. We show that audiovisual adaptation is associated with changes in regional BOLD-responses and fine-scale activity patterns in a widespread network from Heschl's gyrus to dorsolateral prefrontal cortices. Audiovisual recalibration relies on distinct spatial and decisional codes that are expressed with opposite gradients and time courses across the auditory processing hierarchy. Early activity patterns in auditory cortices encode sounds in a continuous space that flexibly adapts to misaligned visual inputs. Later activity patterns in frontoparietal cortices code decisional uncertainty consistent with these spatial transformations. Our findings suggest that regions within the auditory processing hierarchy multiplex spatial and decisional codes to adapt flexibly to the changing sensory statistics in the environment.


Subject(s)
Auditory Cortex , Auditory Perception , Acoustic Stimulation , Adult , Auditory Cortex/physiology , Auditory Perception/physiology , Brain Mapping , Humans , Magnetic Resonance Imaging , Photic Stimulation , Psychophysics , Visual Perception/physiology
10.
Nat Hum Behav ; 6(5): 732-742, 2022 05.
Article in English | MEDLINE | ID: mdl-35210592

ABSTRACT

An intriguing notion in cognitive neuroscience posits that alpha oscillations mould how the brain parses the constant influx of sensory signals into discrete perceptual events. Yet, the evidence is controversial and the underlying neural mechanism unclear. Further, it is unknown whether alpha oscillations influence observers' perceptual sensitivity (that is, temporal resolution) or their top-down biases to bind signals within and across the senses. Combining electroencephalography, psychophysics and signal detection theory, this multi-day study rigorously assessed the impact of alpha frequency on temporal binding of signals within and across the senses. In a series of two-flash discrimination experiments, 20 human observers were presented with one or two flashes together with none, one or two sounds. Our results provide robust evidence that pre-stimulus alpha frequency as a dynamic neural state and an individual's trait index do not influence observers' perceptual sensitivity or bias for two-flash discrimination in any of the three sensory contexts. These results challenge the notion that alpha oscillations have a profound impact on how observers parse sensory inputs into discrete perceptual events.


Subject(s)
Electroencephalography , Visual Perception , Brain , Humans , Photic Stimulation/methods , Sensation
12.
PLoS Biol ; 19(11): e3001465, 2021 11.
Article in English | MEDLINE | ID: mdl-34793436

ABSTRACT

To form a percept of the multisensory world, the brain needs to integrate signals from common sources weighted by their reliabilities and segregate those from independent sources. Previously, we have shown that anterior parietal cortices combine sensory signals into representations that take into account the signals' causal structure (i.e., common versus independent sources) and their sensory reliabilities as predicted by Bayesian causal inference. The current study asks to what extent and how attentional mechanisms can actively control how sensory signals are combined for perceptual inference. In a pre- and postcueing paradigm, we presented observers with audiovisual signals at variable spatial disparities. Observers were precued to attend to auditory or visual modalities prior to stimulus presentation and postcued to report their perceived auditory or visual location. Combining psychophysics, functional magnetic resonance imaging (fMRI), and Bayesian modelling, we demonstrate that the brain moulds multisensory inference via two distinct mechanisms. Prestimulus attention to vision enhances the reliability and influence of visual inputs on spatial representations in visual and posterior parietal cortices. Poststimulus report determines how parietal cortices flexibly combine sensory estimates into spatial representations consistent with Bayesian causal inference. Our results show that distinct neural mechanisms control how signals are combined for perceptual inference at different levels of the cortical hierarchy.


Subject(s)
Attention/physiology , Cerebral Cortex/physiology , Sensation/physiology , Adolescent , Adult , Bayes Theorem , Female , Humans , Magnetic Resonance Imaging , Male , Multivariate Analysis , Oxygen/blood , Time Factors , Young Adult
13.
Trends Cogn Sci ; 25(12): 1013-1014, 2021 12.
Article in English | MEDLINE | ID: mdl-34561193

ABSTRACT

Perception requires the brain to infer whether signals arise from common causes and should hence be integrated or else be treated independently. Rideaux et al. show that a feedforward network can perform causal inference in visuovestibular motion estimation by reading out activity from neurons tuned to congruent and opposite directions.


Subject(s)
Brain , Visual Perception , Brain/physiology , Humans , Neurons , Visual Perception/physiology
14.
Front Hum Neurosci ; 15: 702768, 2021.
Article in English | MEDLINE | ID: mdl-34456697

ABSTRACT

Language comprehension relies on integrating words into progressively more complex structures, like phrases and sentences. This hierarchical structure-building is reflected in rhythmic neural activity across multiple timescales in E/MEG in healthy, awake participants. However, recent studies have shown evidence for this "cortical tracking" of higher-level linguistic structures also in a proportion of unresponsive patients. What does this tell us about these patients' residual levels of cognition and consciousness? Must the listener direct their attention toward higher level speech structures to exhibit cortical tracking, and would selective attention across levels of the hierarchy influence the expression of these rhythms? We investigated these questions in an EEG study of 72 healthy human volunteers listening to streams of monosyllabic isochronous English words that were either unrelated (scrambled condition) or composed of four-word-sequences building meaningful sentences (sentential condition). Importantly, there were no physical cues between four-word-sentences. Rather, boundaries were marked by syntactic structure and thematic role assignment. Participants were divided into three attention groups: from passive listening (passive group) to attending to individual words (word group) or sentences (sentence group). The passive and word groups were initially naïve to the sentential stimulus structure, while the sentence group was not. We found significant tracking at word- and sentence rate across all three groups, with sentence tracking linked to left middle temporal gyrus and right superior temporal gyrus. Goal-directed attention to words did not enhance word-rate-tracking, suggesting that word tracking here reflects largely automatic mechanisms, as was shown for tracking at the syllable-rate before. Importantly, goal-directed attention to sentences relative to words significantly increased sentence-rate-tracking over left inferior frontal gyrus. This attentional modulation of rhythmic EEG activity at the sentential rate highlights the role of attention in integrating individual words into complex linguistic structures. Nevertheless, given the presence of high-level cortical tracking under conditions of lower attentional effort, our findings underline the suitability of the paradigm in its clinical application in patients after brain injury. The neural dissociation between passive tracking of sentences and directed attention to sentences provides a potential means to further characterise the cognitive state of each unresponsive patient.

15.
Sci Rep ; 11(1): 10832, 2021 05 25.
Article in English | MEDLINE | ID: mdl-34035358

ABSTRACT

Information integration is considered a hallmark of human consciousness. Recent research has challenged this tenet by showing multisensory interactions in the absence of awareness. This psychophysics study assessed the impact of spatial and semantic correspondences on audiovisual binding in the presence and absence of visual awareness by combining forward-backward masking with spatial ventriloquism. Observers were presented with object pictures and synchronous sounds that were spatially and/or semantically congruent or incongruent. On each trial observers located the sound, identified the picture and rated the picture's visibility. We observed a robust ventriloquist effect for subjectively visible and invisible pictures indicating that pictures that evade our perceptual awareness influence where we perceive sounds. Critically, semantic congruency enhanced these visual biases on perceived sound location only when the picture entered observers' awareness. Our results demonstrate that crossmodal influences operating from vision to audition and vice versa are interactively controlled by spatial and semantic congruency in the presence of awareness. However, when visual processing is disrupted by masking procedures audiovisual interactions no longer depend on semantic correspondences.


Subject(s)
Auditory Perception/physiology , Space Perception/physiology , Visual Perception/physiology , Acoustic Stimulation , Adult , Female , Humans , Male , Middle Aged , Psychoacoustics , Semantics , Young Adult
16.
Annu Rev Neurosci ; 44: 449-473, 2021 07 08.
Article in English | MEDLINE | ID: mdl-33882258

ABSTRACT

Adaptive behavior in a complex, dynamic, and multisensory world poses some of the most fundamental computational challenges for the brain, notably inference, decision-making, learning, binding, and attention. We first discuss how the brain integrates sensory signals from the same source to support perceptual inference and decision-making by weighting them according to their momentary sensory uncertainties. We then show how observers solve the binding or causal inference problem-deciding whether signals come from common causes and should hence be integrated or else be treated independently. Next, we describe the multifarious interplay between multisensory processing and attention. We argue that attentional mechanisms are crucial to compute approximate solutions to the binding problem in naturalistic environments when complex time-varying signals arise from myriad causes. Finally, we review how the brain dynamically adapts multisensory processing to a changing world across multiple timescales.


Subject(s)
Attention , Auditory Perception , Brain , Learning , Visual Perception
17.
Cortex ; 138: 1-23, 2021 05.
Article in English | MEDLINE | ID: mdl-33676086

ABSTRACT

The processing of multisensory signals is crucial for effective interaction with the environment, but our ability to perform this vital function changes as we age. In the first part of this review, we summarise existing research into the effects of healthy ageing on multisensory integration. We note that age differences vary substantially with the paradigms and stimuli used: older adults often receive at least as much benefit (to both accuracy and response times) as younger controls from congruent multisensory stimuli, but are also consistently more negatively impacted by the presence of intersensory conflict. In the second part, we outline a normative Bayesian framework that provides a principled and computationally informed perspective on the key ingredients involved in multisensory perception, and how these are affected by ageing. Applying this framework to the existing literature, we conclude that changes to sensory reliability, prior expectations (together with attentional control), and decisional strategies all contribute to the age differences observed. However, we find no compelling evidence of any age-related changes to the basic inference mechanisms involved in multisensory perception.


Subject(s)
Auditory Perception , Visual Perception , Acoustic Stimulation , Aged , Bayes Theorem , Humans , Photic Stimulation , Reproducibility of Results
18.
Ann Neurol ; 89(4): 646-656, 2021 04.
Article in English | MEDLINE | ID: mdl-33368496

ABSTRACT

OBJECTIVE: Patients with traumatic brain injury who fail to obey commands after sedation-washout pose one of the most significant challenges for neurological prognostication. Reducing prognostic uncertainty will lead to more appropriate care decisions and ensure provision of limited rehabilitation resources to those most likely to benefit. Bedside markers of covert residual cognition, including speech comprehension, may reduce this uncertainty. METHODS: We recruited 28 patients with acute traumatic brain injury who were 2 to 7 days sedation-free and failed to obey commands. Patients heard streams of isochronous monosyllabic words that built meaningful phrases and sentences while their brain activity via electroencephalography (EEG) was recorded. In healthy individuals, EEG activity only synchronizes with the rhythm of phrases and sentences when listeners consciously comprehend the speech. This approach therefore provides a measure of residual speech comprehension in unresponsive patients. RESULTS: Seventeen and 16 patients were available for assessment with the Glasgow Outcome Scale Extended (GOSE) at 3 months and 6 months, respectively. Outcome significantly correlated with the strength of patients' acute cortical tracking of phrases and sentences (r > 0.6, p < 0.007), quantified by inter-trial phase coherence. Linear regressions revealed that the strength of this comprehension response (beta = 0.603, p = 0.006) significantly improved the accuracy of prognoses relative to clinical characteristics alone (eg, Glasgow Coma Scale [GCS], computed tomography [CT] grade). INTERPRETATION: A simple, passive, auditory EEG protocol improves prognostic accuracy in a critical period of clinical decision making. Unlike other approaches to probing covert cognition for prognostication, this approach is entirely passive and therefore less susceptible to cognitive deficits, increasing the number of patients who may benefit. ANN NEUROL 2021;89:646-656.


Subject(s)
Brain Death/diagnosis , Comprehension , Speech , Adult , Aged , Aged, 80 and over , Brain Death/diagnostic imaging , Brain Injuries, Traumatic/diagnosis , Brain Injuries, Traumatic/psychology , Cerebral Cortex/physiopathology , Electroencephalography , Female , Glasgow Outcome Scale , Humans , Linear Models , Male , Middle Aged , Predictive Value of Tests , Prognosis , Tomography, X-Ray Computed
19.
Elife ; 92020 12 15.
Article in English | MEDLINE | ID: mdl-33319749

ABSTRACT

To form a more reliable percept of the environment, the brain needs to estimate its own sensory uncertainty. Current theories of perceptual inference assume that the brain computes sensory uncertainty instantaneously and independently for each stimulus. We evaluated this assumption in four psychophysical experiments, in which human observers localized auditory signals that were presented synchronously with spatially disparate visual signals. Critically, the visual noise changed dynamically over time continuously or with intermittent jumps. Our results show that observers integrate audiovisual inputs weighted by sensory uncertainty estimates that combine information from past and current signals consistent with an optimal Bayesian learner that can be approximated by exponential discounting. Our results challenge leading models of perceptual inference where sensory uncertainty estimates depend only on the current stimulus. They demonstrate that the brain capitalizes on the temporal dynamics of the external world and estimates sensory uncertainty by combining past experiences with new incoming sensory signals.


Subject(s)
Auditory Perception/physiology , Brain/physiology , Uncertainty , Visual Perception/physiology , Adolescent , Adult , Bayes Theorem , Female , Humans , Male , Middle Aged , Noise , Psychophysics , Young Adult
20.
J Vis ; 20(8): 1, 2020 08 03.
Article in English | MEDLINE | ID: mdl-32744617

ABSTRACT

In our natural environment, the brain needs to combine signals from multiple sensory modalities into a coherent percept. Whereas spatial attention guides perceptual decisions by prioritizing processing of signals that are task-relevant, spatial expectations encode the probability of signals over space. Previous studies have shown that behavioral effects of spatial attention generalize across sensory modalities. However, because they manipulated spatial attention as signal probability over space, these studies could not dissociate attention and expectation or assess their interaction. In two experiments, we orthogonally manipulated spatial attention (i.e., task-relevance) and expectation (i.e., signal probability) selectively in one sensory modality (i.e., primary modality) (experiment 1: audition, experiment 2: vision) and assessed their effects on primary and secondary sensory modalities in which attention and expectation were held constant. Our results show behavioral effects of spatial attention that are comparable for audition and vision as primary modalities; however, signal probabilities were learned more slowly in audition, so that spatial expectations were formed later in audition than vision. Critically, when these differences in learning between audition and vision were accounted for, both spatial attention and expectation affected responses more strongly in the primary modality in which they were manipulated and generalized to the secondary modality only in an attenuated fashion. Collectively, our results suggest that both spatial attention and expectation rely on modality-specific and multisensory mechanisms.


Subject(s)
Attention/physiology , Auditory Perception/physiology , Pattern Recognition, Visual/physiology , Psychomotor Performance/physiology , Space Perception/physiology , Adult , Female , Humans , Male , Motivation , Uncertainty
SELECTION OF CITATIONS
SEARCH DETAIL
...