Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 12 de 12
Filter
Add more filters










Publication year range
1.
J Cogn Neurosci ; 30(11): 1590-1605, 2018 11.
Article in English | MEDLINE | ID: mdl-30125217

ABSTRACT

Object recognition requires dynamic transformations of low-level visual inputs to complex semantic representations. Although this process depends on the ventral visual pathway, we lack an incremental account from low-level inputs to semantic representations and the mechanistic details of these dynamics. Here we combine computational models of vision with semantics and test the output of the incremental model against patterns of neural oscillations recorded with magnetoencephalography in humans. Representational similarity analysis showed visual information was represented in low-frequency activity throughout the ventral visual pathway, and semantic information was represented in theta activity. Furthermore, directed connectivity showed visual information travels through feedforward connections, whereas visual information is transformed into semantic representations through feedforward and feedback activity, centered on the anterior temporal lobe. Our research highlights that the complex transformations between visual and semantic information is driven by feedforward and recurrent dynamics resulting in object-specific semantics.


Subject(s)
Pattern Recognition, Visual/physiology , Photic Stimulation/methods , Semantics , Theta Rhythm/physiology , Visual Pathways/diagnostic imaging , Visual Pathways/physiology , Computational Biology/methods , Female , Humans , Magnetic Resonance Imaging/methods , Magnetoencephalography/methods , Male , Random Allocation
2.
Sci Rep ; 8(1): 10636, 2018 Jul 13.
Article in English | MEDLINE | ID: mdl-30006530

ABSTRACT

Recognising an object involves rapid visual processing and activation of semantic knowledge about the object, but how visual processing activates and interacts with semantic representations remains unclear. Cognitive neuroscience research has shown that while visual processing involves posterior regions along the ventral stream, object meaning involves more anterior regions, especially perirhinal cortex. Here we investigate visuo-semantic processing by combining a deep neural network model of vision with an attractor network model of semantics, such that visual information maps onto object meanings represented as activation patterns across features. In the combined model, concept activation is driven by visual input and co-occurrence of semantic features, consistent with neurocognitive accounts. We tested the model's ability to explain fMRI data where participants named objects. Visual layers explained activation patterns in early visual cortex, whereas pattern-information in perirhinal cortex was best explained by later stages of the attractor network, when detailed semantic representations are activated. Posterior ventral temporal cortex was best explained by intermediate stages corresponding to initial semantic processing, when visual information has the greatest influence on the emerging semantic representation. These results provide proof of principle of how a mechanistic model of combined visuo-semantic processing can account for pattern-information in the ventral stream.


Subject(s)
Models, Neurological , Pattern Recognition, Visual/physiology , Perirhinal Cortex/physiology , Visual Cortex/physiology , Visual Pathways/physiology , Brain Mapping/methods , Female , Humans , Image Processing, Computer-Assisted/methods , Magnetic Resonance Imaging/methods , Male , Perirhinal Cortex/diagnostic imaging , Semantics , Visual Cortex/diagnostic imaging , Visual Pathways/diagnostic imaging
3.
Lang Cogn Neurosci ; 32(2): 221-235, 2017 Feb 07.
Article in English | MEDLINE | ID: mdl-28164141

ABSTRACT

As spoken language unfolds over time the speech input transiently activates multiple candidates at different levels of the system - phonological, lexical, and syntactic - which in turn leads to short-lived between-candidate competition. In an fMRI study, we investigated how different kinds of linguistic competition may be modulated by the presence or absence of a prior context (Tyler 1984; Tyler et al. 2008). We found significant effects of lexico-phonological competition for isolated words, but not for words in short phrases, with high competition yielding greater activation in left inferior frontal gyrus (LIFG) and posterior temporal regions. This suggests that phrasal contexts reduce lexico-phonological competition by eliminating form-class inconsistent cohort candidates. A corpus-derived measure of lexico-syntactic competition was associated with greater activation in LIFG for verbs in phrases, but not for isolated verbs, indicating that lexico-syntactic information is boosted by the phrasal context. Together, these findings indicate that LIFG plays a general role in resolving different kinds of linguistic competition.

4.
J Neurosci ; 37(5): 1312-1319, 2017 02 01.
Article in English | MEDLINE | ID: mdl-28028201

ABSTRACT

Comprehending speech involves the rapid and optimally efficient mapping from sound to meaning. Influential cognitive models of spoken word recognition (Marslen-Wilson and Welsh, 1978) propose that the onset of a spoken word initiates a continuous process of activation of the lexical and semantic properties of the word candidates matching the speech input and competition between them, which continues until the point at which the word is differentiated from all other cohort candidates (the uniqueness point, UP). At this point, the word is recognized uniquely and only the target word's semantics are active. Although it is well established that spoken word recognition engages the superior (Rauschecker and Scott, 2009), middle, and inferior (Hickok and Poeppel, 2007) temporal cortices, little is known about the real-time brain activity that underpins the computations and representations that evolve over time during the transformation from speech to meaning. Here, we test for the first time the spatiotemporal dynamics of these processes by collecting MEG data while human participants listened to spoken words. By constructing quantitative models of competition and access to meaning in combination with spatiotemporal searchlight representational similarity analysis (Kriegeskorte et al., 2006) in source space, we were able to test where and when these models produced significant effects. We found early transient effects ∼400 ms before the UP of lexical competition in left supramarginal gyrus, left superior temporal gyrus, left middle temporal gyrus (MTG), and left inferior frontal gyrus (IFG) and of semantic competition in MTG, left angular gyrus, and IFG. After the UP, there were no competitive effects, only target-specific semantic effects in angular gyrus and MTG. SIGNIFICANCE STATEMENT: Understanding spoken words involves complex processes that transform the auditory input into a meaningful interpretation. This effortless transition occurs on millisecond timescales, with remarkable speed and accuracy and without any awareness of the complex computations involved. Here, we reveal the real-time neural dynamics of these processes by collecting data about listeners' brain activity as they hear spoken words. Using novel statistical models of different aspects of the recognition process, we can locate directly which parts of the brain are accessing the stored form and meaning of words and how the competition between different word candidates is resolved neurally in real time. This gives us a uniquely differentiated picture of the neural substrate for the first 500 ms of word recognition.


Subject(s)
Auditory Perception/physiology , Cerebral Cortex/physiology , Comprehension/physiology , Adult , Brain Mapping , Electroencephalography , Female , Humans , Magnetoencephalography , Male , Psychomotor Performance/physiology , Recognition, Psychology/physiology , Sensory Gating/physiology , Sound Localization/physiology , Speech Perception/physiology , Temporal Lobe/physiology , Young Adult
5.
Cogn Sci ; 40(2): 325-50, 2016 Mar.
Article in English | MEDLINE | ID: mdl-26043761

ABSTRACT

Understanding spoken words involves a rapid mapping from speech to conceptual representations. One distributed feature-based conceptual account assumes that the statistical characteristics of concepts' features--the number of concepts they occur in (distinctiveness/sharedness) and likelihood of co-occurrence (correlational strength)--determine conceptual activation. To test these claims, we investigated the role of distinctiveness/sharedness and correlational strength in speech-to-meaning mapping, using a lexical decision task and computational simulations. Responses were faster for concepts with higher sharedness, suggesting that shared features are facilitatory in tasks like lexical decision that require access to them. Correlational strength facilitated responses for slower participants, suggesting a time-sensitive co-occurrence-driven settling mechanism. The computational simulation showed similar effects, with early effects of shared features and later effects of correlational strength. These results support a general-to-specific account of conceptual processing, whereby early activation of shared features is followed by the gradual emergence of a specific target representation.


Subject(s)
Comprehension/physiology , Concept Formation/physiology , Models, Theoretical , Speech/physiology , Adolescent , Adult , Computer Simulation , Humans , Reaction Time , Time Factors , Young Adult
6.
Cereb Cortex ; 25(10): 3602-12, 2015 Oct.
Article in English | MEDLINE | ID: mdl-25209607

ABSTRACT

To respond appropriately to objects, we must process visual inputs rapidly and assign them meaning. This involves highly dynamic, interactive neural processes through which information accumulates and cognitive operations are resolved across multiple time scales. However, there is currently no model of object recognition which provides an integrated account of how visual and semantic information emerge over time; therefore, it remains unknown how and when semantic representations are evoked from visual inputs. Here, we test whether a model of individual objects--based on combining the HMax computational model of vision with semantic-feature information--can account for and predict time-varying neural activity recorded with magnetoencephalography. We show that combining HMax and semantic properties provides a better account of neural object representations compared with the HMax alone, both through model fit and classification performance. Our results show that modeling and classifying individual objects is significantly improved by adding semantic-feature information beyond ∼200 ms. These results provide important insights into the functional properties of visual processing across time.


Subject(s)
Cerebral Cortex/physiology , Pattern Recognition, Visual/physiology , Recognition, Psychology/physiology , Semantics , Adult , Concept Formation/physiology , Female , Humans , Magnetoencephalography , Male , Models, Neurological , Regression Analysis , Young Adult
7.
Behav Res Methods ; 46(4): 1119-27, 2014 Dec.
Article in English | MEDLINE | ID: mdl-24356992

ABSTRACT

Theories of the representation and processing of concepts have been greatly enhanced by models based on information available in semantic property norms. This information relates both to the identity of the features produced in the norms and to their statistical properties. In this article, we introduce a new and large set of property norms that are designed to be a more flexible tool to meet the demands of many different disciplines interested in conceptual knowledge representation, from cognitive psychology to computational linguistics. As well as providing all features listed by 2 or more participants, we also show the considerable linguistic variation that underlies each normalized feature label and the number of participants who generated each variant. Our norms are highly comparable with the largest extant set (McRae, Cree, Seidenberg, & McNorgan, 2005) in terms of the number and distribution of features. In addition, we show how the norms give rise to a coherent category structure. We provide these norms in the hope that the greater detail available in the Centre for Speech, Language and the Brain norms should further promote the development of models of conceptual knowledge. The norms can be downloaded at www.csl.psychol.cam.ac.uk/propertynorms.


Subject(s)
Concept Formation/classification , Language , Semantics , Speech , Adolescent , Adult , Female , Humans , Linguistics , Male , Reference Values , Word Association Tests , Young Adult
8.
J Neurosci ; 33(48): 18906-16, 2013 Nov 27.
Article in English | MEDLINE | ID: mdl-24285896

ABSTRACT

Understanding the meanings of words and objects requires the activation of underlying conceptual representations. Semantic representations are often assumed to be coded such that meaning is evoked regardless of the input modality. However, the extent to which meaning is coded in modality-independent or amodal systems remains controversial. We address this issue in a human fMRI study investigating the neural processing of concepts, presented separately as written words and pictures. Activation maps for each individual word and picture were used as input for searchlight-based multivoxel pattern analyses. Representational similarity analysis was used to identify regions correlating with low-level visual models of the words and objects and the semantic category structure common to both. Common semantic category effects for both modalities were found in a left-lateralized network, including left posterior middle temporal gyrus (LpMTG), left angular gyrus, and left intraparietal sulcus (LIPS), in addition to object- and word-specific semantic processing in ventral temporal cortex and more anterior MTG, respectively. To explore differences in representational content across regions and modalities, we developed novel data-driven analyses, based on k-means clustering of searchlight dissimilarity matrices and seeded correlation analysis. These revealed subtle differences in the representations in semantic-sensitive regions, with representations in LIPS being relatively invariant to stimulus modality and representations in LpMTG being uncorrelated across modality. These results suggest that, although both LpMTG and LIPS are involved in semantic processing, only the functional role of LIPS is the same regardless of the visual input, whereas the functional role of LpMTG differs for words and objects.


Subject(s)
Form Perception/physiology , Reading , Semantics , Visual Perception/physiology , Adult , Brain Mapping , Cluster Analysis , Female , Humans , Image Processing, Computer-Assisted , Magnetic Resonance Imaging , Male , Photic Stimulation , Psycholinguistics , Young Adult
9.
Front Psychol ; 4: 271, 2013.
Article in English | MEDLINE | ID: mdl-23730293

ABSTRACT

The core human capacity of syntactic analysis involves a left hemisphere network involving left inferior frontal gyrus (LIFG) and posterior middle temporal gyrus (LMTG) and the anatomical connections between them. Here we use magnetoencephalography (MEG) to determine the spatio-temporal properties of syntactic computations in this network. Listeners heard spoken sentences containing a local syntactic ambiguity (e.g., "… landing planes …"), at the offset of which they heard a disambiguating verb and decided whether it was an acceptable/unacceptable continuation of the sentence. We charted the time-course of processing and resolving syntactic ambiguity by measuring MEG responses from the onset of each word in the ambiguous phrase and the disambiguating word. We used representational similarity analysis (RSA) to characterize syntactic information represented in the LIFG and left posterior middle temporal gyrus (LpMTG) over time and to investigate their relationship to each other. Testing a variety of lexico-syntactic and ambiguity models against the MEG data, our results suggest early lexico-syntactic responses in the LpMTG and later effects of ambiguity in the LIFG, pointing to a clear differentiation in the functional roles of these two regions. Our results suggest the LpMTG represents and transmits lexical information to the LIFG, which responds to and resolves the ambiguity.

10.
J Cogn Neurosci ; 25(10): 1723-35, 2013 Oct.
Article in English | MEDLINE | ID: mdl-23662861

ABSTRACT

Recognizing an object involves more than just visual analyses; its meaning must also be decoded. Extensive research has shown that processing the visual properties of objects relies on a hierarchically organized stream in ventral occipitotemporal cortex, with increasingly more complex visual features being coded from posterior to anterior sites culminating in the perirhinal cortex (PRC) in the anteromedial temporal lobe (aMTL). The neurobiological principles of the conceptual analysis of objects remain more controversial. Much research has focused on two neural regions-the fusiform gyrus and aMTL, both of which show semantic category differences, but of different types. fMRI studies show category differentiation in the fusiform gyrus, based on clusters of semantically similar objects, whereas category-specific deficits, specifically for living things, are associated with damage to the aMTL. These category-specific deficits for living things have been attributed to problems in differentiating between highly similar objects, a process that involves the PRC. To determine whether the PRC and the fusiform gyri contribute to different aspects of an object's meaning, with differentiation between confusable objects in the PRC and categorization based on object similarity in the fusiform, we carried out an fMRI study of object processing based on a feature-based model that characterizes the degree of semantic similarity and difference between objects and object categories. Participants saw 388 objects for which feature statistic information was available and named the objects at the basic level while undergoing fMRI scanning. After controlling for the effects of visual information, we found that feature statistics that capture similarity between objects formed category clusters in fusiform gyri, such that objects with many shared features (typical of living things) were associated with activity in the lateral fusiform gyri whereas objects with fewer shared features (typical of nonliving things) were associated with activity in the medial fusiform gyri. Significantly, a feature statistic reflecting differentiation between highly similar objects, enabling object-specific representations, was associated with bilateral PRC activity. These results confirm that the statistical characteristics of conceptual object features are coded in the ventral stream, supporting a conceptual feature-based hierarchy, and integrating disparate findings of category responses in fusiform gyri and category deficits in aMTL into a unifying neurocognitive framework.


Subject(s)
Pattern Recognition, Visual/physiology , Recognition, Psychology/physiology , Visual Cortex/physiology , Visual Pathways/physiology , Adult , Brain Mapping , Female , Humans , Image Processing, Computer-Assisted , Magnetic Resonance Imaging , Male , Names , Oxygen/blood , Reaction Time , Visual Cortex/blood supply , Visual Pathways/blood supply , Young Adult
11.
Cognition ; 122(3): 363-74, 2012 Mar.
Article in English | MEDLINE | ID: mdl-22137770

ABSTRACT

Conceptual representations are at the heart of our mental lives, involved in every aspect of cognitive functioning. Despite their centrality, a long-standing debate persists as to how the meanings of concepts are represented and processed. Many accounts agree that the meanings of concrete concepts are represented by their individual features, but disagree about the importance of different feature-based variables: some views stress the importance of the information carried by distinctive features in conceptual processing, others the features which are shared over many concepts, and still others the extent to which features co-occur. We suggest that previously disparate theoretical positions and experimental findings can be unified by an account which claims that task demands determine how concepts are processed in addition to the effects of feature distinctiveness and co-occurrence. We tested these predictions in a basic-level naming task which relies on distinctive feature information (Experiment 1) and a domain decision task which relies on shared feature information (Experiment 2). Both used large-scale regression designs with the same visual objects, and mixed-effects models incorporating participant, session, stimulus-related and feature statistic variables to model the performance. We found that concepts with relatively more distinctive and more highly correlated distinctive relative to shared features facilitated basic-level naming latencies, while concepts with relatively more shared and more highly correlated shared relative to distinctive features speeded domain decisions. These findings demonstrate that the feature statistics of distinctiveness (shared vs. distinctive) and correlational strength, as well as the task demands, determine how concept meaning is processed in the conceptual system.


Subject(s)
Concept Formation/physiology , Visual Perception/physiology , Adolescent , Adult , Attention/physiology , Cognition/physiology , Female , Humans , Male , Psycholinguistics , Semantics
12.
Brain ; 134(Pt 2): 415-31, 2011 Feb.
Article in English | MEDLINE | ID: mdl-21278407

ABSTRACT

For the past 150 years, neurobiological models of language have debated the role of key brain regions in language function. One consistently debated set of issues concern the role of the left inferior frontal gyrus in syntactic processing. Here we combine measures of functional activity, grey matter integrity and performance in patients with left hemisphere damage and healthy participants to ask whether the left inferior frontal gyrus is essential for syntactic processing. In a functional neuroimaging study, participants listened to spoken sentences that either contained a syntactically ambiguous or matched unambiguous phrase. Behavioural data on three tests of syntactic processing were subsequently collected. In controls, syntactic processing co-activated left hemisphere Brodmann areas 45/47 and posterior middle temporal gyrus. Activity in a left parietal cluster was sensitive to working memory demands in both patients and controls. Exploiting the variability in lesion location and performance in the patients, voxel-based correlational analyses showed that tissue integrity and neural activity-primarily in left Brodmann area 45 and posterior middle temporal gyrus-were correlated with preserved syntactic performance, but unlike the controls, patients were insensitive to syntactic preferences, reflecting their syntactic deficit. These results argue for the essential contribution of the left inferior frontal gyrus in syntactic analysis and highlight the functional relationship between left Brodmann area 45 and the left posterior middle temporal gyrus, suggesting that when this relationship breaks down, through damage to either region or to the connections between them, syntactic processing is impaired. On this view, the left inferior frontal gyrus may not itself be specialized for syntactic processing, but plays an essential role in the neural network that carries out syntactic computations.


Subject(s)
Aphasia/physiopathology , Aphasia/psychology , Brain Injury, Chronic/physiopathology , Brain Injury, Chronic/psychology , Frontal Lobe/physiopathology , Functional Laterality/physiology , Adult , Aged , Aphasia/pathology , Brain Injury, Chronic/pathology , Brain Mapping/methods , Frontal Lobe/pathology , Humans , Magnetic Resonance Imaging/methods , Male , Middle Aged , Psychomotor Performance/physiology , Speech Perception/physiology , Temporal Lobe/physiopathology
SELECTION OF CITATIONS
SEARCH DETAIL
...