Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 14 de 14
Filter
Add more filters










Publication year range
1.
J Acoust Soc Am ; 152(6): 3728, 2022 12.
Article in English | MEDLINE | ID: mdl-36586841

ABSTRACT

Research on speech categorization and phoneme recognition has relied heavily on tasks in which participants listen to stimuli from a speech continuum and are asked to either classify each stimulus (identification) or discriminate between them (discrimination). Such tasks rest on assumptions about how perception maps onto discrete responses that have not been thoroughly investigated. Here, we identify critical challenges in the link between these tasks and theories of speech categorization. In particular, we show that patterns that have traditionally been linked to categorical perception could arise despite continuous underlying perception and that patterns that run counter to categorical perception could arise despite underlying categorical perception. We describe an alternative measure of speech perception using a visual analog scale that better differentiates between processes at play in speech categorization, and we review some recent findings that show how this task can be used to better inform our theories.


Subject(s)
Speech Perception , Speech , Humans , Auditory Perception , Speech Perception/physiology , Recognition, Psychology , Phonetics
2.
J Exp Psychol Learn Mem Cogn ; 48(3): 394-415, 2022 Mar.
Article in English | MEDLINE | ID: mdl-35389728

ABSTRACT

Does saying a novel word help to recognize it later? Previous research on the effect of production on this aspect of word learning is inconclusive, as both facilitatory and detrimental effects of production are reported. In a set of three experiments, we sought to reconcile the seemingly contrasting findings by disentangling the production from other effects. In Experiment 1, participants learned eight new words and their visual referents. On each trial, participants heard a novel word twice: either (a) by hearing the same speaker produce it twice (Perception-Only condition) or (b) by first hearing the speaker once and then producing it themselves (Production condition). At test, participants saw two pictures while hearing a novel word and were asked to choose its correct referent. Experiment 2 was identical to Experiment 1, except that in the Perception-Only condition each word was spoken by 2 different speakers (equalizing talker variability between conditions). Experiment 3 was identical to Experiment 2, but at test words were spoken by a novel speaker to assess generalizability of the effect. Accuracy, reaction time, and eye-movements to the target image were collected. Production had a facilitatory effect during early stages of learning (after short training), but its effect became detrimental after additional training. The results help to reconcile conflicting findings regarding the role of production on word learning. This work is relevant to a wide range of research on human learning in showing that the same factor may play a different role at different stages of learning. (PsycInfo Database Record (c) 2022 APA, all rights reserved).


Subject(s)
Speech Perception , Humans , Learning , Reaction Time , Verbal Learning
3.
Brain Lang ; 223: 105031, 2021 12.
Article in English | MEDLINE | ID: mdl-34628259

ABSTRACT

Listeners generally categorize speech sounds in a gradient manner. However, recent work, using a visual analogue scaling (VAS) task, suggests that some listeners show more categorical performance, leading to less flexible cue integration and poorer recovery from misperceptions (Kapnoula et al., 2017, 2021). We asked how individual differences in speech gradiency can be reconciled with the well-established gradiency in the modal listener, showing how VAS performance relates to both Visual World Paradigm and EEG measures of gradiency. We also investigated three potential sources of these individual differences: inhibitory control; lexical inhibition; and early cue encoding. We used the N1 ERP component to track pre-categorical encoding of Voice Onset Time (VOT). The N1 linearly tracked VOT, reflecting a fundamentally gradient speech perception; however, for less gradient listeners, this linearity was disrupted near the boundary. Thus, while all listeners are gradient, they may show idiosyncratic encoding of specific cues, affecting downstream processing.


Subject(s)
Speech Perception , Voice , Cues , Eye-Tracking Technology , Humans , Phonetics , Speech , Speech Perception/physiology
4.
Front Psychol ; 12: 689052, 2021.
Article in English | MEDLINE | ID: mdl-34305748

ABSTRACT

The examination of how words are learned can offer valuable insights into the nature of lexical representations. For example, a common assessment of novel word learning is based on its ability to interfere with other words; given that words are known to compete with each other (Luce and Pisoni, 1998; Dahan et al., 2001), we can use the capacity of a novel word to interfere with the activation of other lexical representations as a measure of the degree to which it is integrated into the mental lexicon (Leach and Samuel, 2007). This measure allows us to assess novel word learning in L1 or L2, but also the degree to which representations from the two lexica interact with each other (Marian and Spivey, 2003). Despite the somewhat independent lines of research on L1 and L2 word learning, common patterns emerge across the two literatures (Lindsay and Gaskell, 2010; Palma and Titone, 2020). In both cases, lexicalization appears to follow a similar trajectory. In L1, newly encoded words often fail at first to engage in competition with known words, but they do so later, after they have been better integrated into the mental lexicon (Gaskell and Dumay, 2003; Dumay and Gaskell, 2012; Bakker et al., 2014). Similarly, L2 words generally have a facilitatory effect, which can, however, become inhibitory in the case of more robust (high-frequency) lexical representations. Despite the similar pattern, L1 lexicalization is described in terms of inter-lexical connections (Leach and Samuel, 2007), leading to more automatic processing (McMurray et al., 2016); whereas in L2 word learning, lack of lexical inhibition is attributed to less robust (i.e., fuzzy) L2 lexical representations. Here, I point to these similarities and I use them to argue that a common mechanism may underlie similar patterns across the two literatures.

5.
J Exp Psychol Hum Percept Perform ; 47(4): 578-595, 2021 04.
Article in English | MEDLINE | ID: mdl-33983791

ABSTRACT

Listeners activate speech-sound categories in a gradient way, and this information is maintained and affects activation of items at higher levels of processing (McMurray et al., 2002; Toscano et al., 2010). Recent findings by Kapnoula et al. (2017) suggest that the degree to which listeners maintain within-category information varies across individuals. Here we assessed the consequences of this gradiency for speech perception. To test this, we collected a measure of gradiency for different listeners using the visual analogue scaling (VAS) task used by Kapnoula et al. (2017). We also collected 2 independent measures of performance in speech perception: a visual world paradigm (VWP) task measuring participants' ability to recover from lexical garden paths (McMurray et al., 2009) and a speech-perception task measuring participants' perception of isolated words in noise. Our results show that categorization gradiency does not predict participants' performance in the speech-in-noise task. However, higher gradiency predicted higher likelihood of recovery from temporarily misleading information presented in the VWP task. These results suggest that gradient activation of speech sound categories is helpful when listeners need to reconsider their initial interpretation of the input, making them more efficient in recovering from errors. (PsycInfo Database Record (c) 2021 APA, all rights reserved).


Subject(s)
Speech Perception , Speech , Humans , Noise , Phonetics
6.
Brain Lang ; 211: 104875, 2020 12.
Article in English | MEDLINE | ID: mdl-33086178

ABSTRACT

Understanding spoken language requires analysis of the rapidly unfolding speech signal at multiple levels: acoustic, phonological, and semantic. However, there is not yet a comprehensive picture of how these levels relate. We recorded electroencephalography (EEG) while listeners (N = 31) heard sentences in which we manipulated acoustic ambiguity (e.g., a bees/peas continuum) and sentential expectations (e.g., Honey is made by bees). EEG was analyzed with a mixed effects model over time to quantify how language processing cascades proceed on a millisecond-by-millisecond basis. Our results indicate: (1) perceptual processing and memory for fine-grained acoustics is preserved in brain activity for up to 900 msec; (2) contextual analysis begins early and is graded with respect to the acoustic signal; and (3) top-down predictions influence perceptual processing in some cases, however, these predictions are available simultaneously with the veridical signal. These mechanistic insights provide a basis for a better understanding of the cortical language network.


Subject(s)
Acoustic Stimulation/methods , Comprehension/physiology , Electroencephalography/methods , Language , Motivation/physiology , Speech Perception/physiology , Adult , Auditory Perception/physiology , Female , Humans , Male , Semantics
7.
J Neurosci ; 40(5): 1053-1065, 2020 01 29.
Article in English | MEDLINE | ID: mdl-31889007

ABSTRACT

Lip-reading is crucial for understanding speech in challenging conditions. But how the brain extracts meaning from, silent, visual speech is still under debate. Lip-reading in silence activates the auditory cortices, but it is not known whether such activation reflects immediate synthesis of the corresponding auditory stimulus or imagery of unrelated sounds. To disentangle these possibilities, we used magnetoencephalography to evaluate how cortical activity in 28 healthy adult humans (17 females) entrained to the auditory speech envelope and lip movements (mouth opening) when listening to a spoken story without visual input (audio-only), and when seeing a silent video of a speaker articulating another story (video-only). In video-only, auditory cortical activity entrained to the absent auditory signal at frequencies <1 Hz more than to the seen lip movements. This entrainment process was characterized by an auditory-speech-to-brain delay of ∼70 ms in the left hemisphere, compared with ∼20 ms in audio-only. Entrainment to mouth opening was found in the right angular gyrus at <1 Hz, and in early visual cortices at 1-8 Hz. These findings demonstrate that the brain can use a silent lip-read signal to synthesize a coarse-grained auditory speech representation in early auditory cortices. Our data indicate the following underlying oscillatory mechanism: seeing lip movements first modulates neuronal activity in early visual cortices at frequencies that match articulatory lip movements; the right angular gyrus then extracts slower features of lip movements, mapping them onto the corresponding speech sound features; this information is fed to auditory cortices, most likely facilitating speech parsing.SIGNIFICANCE STATEMENT Lip-reading consists in decoding speech based on visual information derived from observation of a speaker's articulatory facial gestures. Lip-reading is known to improve auditory speech understanding, especially when speech is degraded. Interestingly, lip-reading in silence still activates the auditory cortices, even when participants do not know what the absent auditory signal should be. However, it was uncertain what such activation reflected. Here, using magnetoencephalographic recordings, we demonstrate that it reflects fast synthesis of the auditory stimulus rather than mental imagery of unrelated, speech or non-speech, sounds. Our results also shed light on the oscillatory dynamics underlying lip-reading.


Subject(s)
Auditory Cortex/physiology , Lipreading , Speech Perception/physiology , Acoustic Stimulation , Female , Humans , Magnetoencephalography , Male , Pattern Recognition, Visual/physiology , Sound Spectrography , Young Adult
8.
Parkinsonism Relat Disord ; 63: 46-53, 2019 06.
Article in English | MEDLINE | ID: mdl-30871801

ABSTRACT

INTRODUCTION: Deep brain stimulation (DBS) of the subthalamic nucleus (STN) is an effective treatment for limb motor symptoms in Parkinson's disease (PD); however, its effect on vocal motor function has yielded conflicted and highly variable results. The present study investigated the effects of STN-DBS on the mechanisms of vocal production and motor control. METHODS: A total of 10 PD subjects with bilateral STN-DBS implantation were tested with DBS ON and OFF while they performed steady vowel vocalizations and received randomized upward or downward pitch-shift stimuli (±100 cents) in their voice auditory feedback. RESULTS: Data showed that the magnitude of vocal compensation responses to pitch-shift stimuli was significantly attenuated during DBS ON vs. OFF (p = 0.012). This effect was direction-specific and was only observed when subjects raised their voice fundamental frequency (F0) in the opposite direction to downward stimuli (p = 0.019). In addition, we found that voice F0 perturbation (i.e. jitter) was significantly reduced during DBS ON vs. OFF (p = 0.022), and this DBS-induced modulation was positively correlated with the attenuation of vocal compensation responses to downward pitch-shift stimuli (r = +0.57, p = 0.028). CONCLUSIONS: These findings provide the first data supporting the role of STN in vocal F0 motor control in response to altered auditory feedback. The DBS-induced attenuation of vocal compensation responses may result from increased inhibitory effects of the subcortical hyperdirect (fronto-subthalamic) pathways on the vocal motor cortex, which can help stabilize voice F0 and ameliorate vocal motor symptoms by impeding PD subjects' abnormal (i.e. overshooting) vocal responses to alterations in the auditory feedback.


Subject(s)
Deep Brain Stimulation , Parkinson Disease/therapy , Voice/physiology , Aged , Feedback, Sensory/physiology , Female , Humans , Middle Aged , Motor Cortex/physiopathology , Parkinson Disease/physiopathology , Subthalamic Nucleus/physiology
9.
J Exp Psychol Hum Percept Perform ; 43(9): 1594-1611, 2017 Sep.
Article in English | MEDLINE | ID: mdl-28406683

ABSTRACT

During spoken language comprehension listeners transform continuous acoustic cues into categories (e.g., /b/ and /p/). While long-standing research suggests that phonetic categories are activated in a gradient way, there are also clear individual differences in that more gradient categorization has been linked to various communication impairments such as dyslexia and specific language impairments (Joanisse, Manis, Keating, & Seidenberg, 2000; López-Zamora, Luque, Álvarez, & Cobos, 2012; Serniclaes, Van Heghe, Mousty, Carré, & Sprenger-Charolles, 2004; Werker & Tees, 1987). Crucially, most studies have used 2-alternative forced choice (2AFC) tasks to measure the sharpness of between-category boundaries. Here we propose an alternative paradigm that allows us to measure categorization gradiency in a more direct way. Furthermore, we follow an individual differences approach to (a) link this measure of gradiency to multiple cue integration, (b) explore its relationship to a set of other cognitive processes, and (c) evaluate its role in individuals' ability to perceive speech in noise. Our results provide validation for this new method of assessing phoneme categorization gradiency and offer preliminary insights into how different aspects of speech perception may be linked to each other and to more general cognitive processes. (PsycINFO Database Record


Subject(s)
Cues , Executive Function/physiology , Individuality , Inhibition, Psychological , Memory, Short-Term/physiology , Psycholinguistics , Speech Perception/physiology , Adult , Humans , Psycholinguistics/methods , Young Adult
10.
Neuroimage ; 128: 328-341, 2016 Mar.
Article in English | MEDLINE | ID: mdl-26806289

ABSTRACT

In this study predictions of the dual-route cascaded (DRC) model of word reading were tested using fMRI. Specifically, patterns of co-localization were investigated: (a) between pseudoword length effects and a pseudowords vs. fixation contrast, to reveal the sublexical grapho-phonemic conversion (GPC) system; and (b) between word frequency effects and a words vs. pseudowords contrast, to reveal the orthographic and phonological lexicon. Forty four native speakers of Greek were scanned at 3T in an event-related lexical decision task with three event types: (a) 150 words in which frequency, length, bigram and syllable frequency, neighborhood, and orthographic consistency were decorrelated; (b) 150 matched pseudowords; and (c) fixation. Whole-brain analysis failed to reveal the predicted co-localizations. Further analysis with participant-specific regions of interest defined within masks from the group contrasts revealed length effects in left inferior parietal cortex and frequency effects in the left middle temporal gyrus. These findings could be interpreted as partially consistent with the existence of the GPC system and phonological lexicon of the model, respectively. However, there was no evidence in support of an orthographic lexicon, weakening overall support for the model. The results are discussed with respect to the prospect of using neuroimaging in cognitive model evaluation.


Subject(s)
Brain/physiology , Pattern Recognition, Visual/physiology , Reading , Adult , Female , Humans , Image Processing, Computer-Assisted , Magnetic Resonance Imaging , Male , Middle Aged , Recognition, Psychology/physiology , Young Adult
11.
J Exp Psychol Learn Mem Cogn ; 42(4): 542-65, 2016 Apr.
Article in English | MEDLINE | ID: mdl-26436633

ABSTRACT

Effects of lexical and sublexical variables on visual word recognition are often treated as homogeneous across participants and stable over time. In this study, we examine the modulation of frequency, length, syllable and bigram frequency, orthographic neighborhood, and graphophonemic consistency effects by (a) individual differences, and (b) item repetition. A group of 129 participants performed lexical decision and naming, in counterbalanced order, using a set of 150 Greek words in which these variables were decorrelated. Frequency, length, and syllable frequency effects were reduced by a preceding task. Length effects were inversely related to years of education. Neighborhood effects depended on the metric used. There were no significant effects or interactions of bigram frequency or consistency. The results suggest that exposure to a word causes transient effects that may cumulatively develop into permanent individual differences. Models of word recognition must incorporate item-specific learning to account for these findings.


Subject(s)
Memory, Long-Term , Memory, Short-Term , Pattern Recognition, Visual , Psycholinguistics , Reading , Adolescent , Adult , Decision Making , Educational Status , Female , Humans , Language Tests , Male , Reaction Time , Young Adult
12.
J Exp Psychol Gen ; 145(1): 8-30, 2016 Jan.
Article in English | MEDLINE | ID: mdl-26709587

ABSTRACT

Language learning is generally described as a problem of acquiring new information (e.g., new words). However, equally important are changes in how the system processes known information. For example, a wealth of studies has suggested dramatic changes over development in how efficiently children recognize familiar words, but it is unknown what kind of experience-dependent mechanisms of plasticity give rise to such changes in real-time processing. We examined the plasticity of the language processing system by testing whether a fundamental aspect of spoken word recognition, lexical interference, can be altered by experience. Adult participants were trained on a set of familiar words over a series of 4 tasks. In the high-competition (HC) condition, tasks were designed to encourage coactivation of similar words (e.g., net and neck) and to require listeners to resolve this competition. Tasks were similar in the low-competition (LC) condition, but did not enhance this competition. Immediately after training, interlexical interference was tested using a visual world paradigm task. Participants in the HC group resolved interference to a fuller degree than those in the LC group, demonstrating that experience can shape the way competition between words is resolved. TRACE simulations showed that the observed late differences in the pattern of interference resolution can be attributed to differences in the strength of lexical inhibition. These findings inform cognitive models in many domains that involve competition/interference processes, and suggest an experience-dependent mechanism of plasticity that may underlie longer term changes in processing efficiency associated with both typical and atypical development.


Subject(s)
Attention , Inhibition, Psychological , Phonetics , Practice, Psychological , Reading , Recognition, Psychology , Semantics , Speech Perception , Verbal Learning , Adult , Child , Female , Humans , Male , Paired-Associate Learning , Reaction Time
13.
Psychon Bull Rev ; 23(2): 491-9, 2016 Apr.
Article in English | MEDLINE | ID: mdl-26202702

ABSTRACT

A hotly debated question in word learning concerns the conditions under which newly learned words compete or interfere with familiar words during spoken word recognition. This has recently been described as a key marker of the integration of a new word into the lexicon and was thought to require consolidation Dumay & Gaskell, (Psychological Science, 18, 35-39, 2007; Gaskell & Dumay, Cognition, 89, 105-132, 2003). Recently, however, Kapnoula, Packard, Gupta, and McMurray, (Cognition, 134, 85-99, 2015) showed that interference can be observed immediately after a word is first learned, implying very rapid integration of new words into the lexicon. It is an open question whether these kinds of effects derive from episodic traces of novel words or from more abstract and lexicalized representations. Here we addressed this question by testing inhibition for newly learned words using training and test stimuli presented in different talker voices. During training, participants were exposed to a set of nonwords spoken by a female speaker. Immediately after training, we assessed the ability of the novel word forms to inhibit familiar words, using a variant of the visual world paradigm. Crucially, the test items were produced by a male speaker. An analysis of fixations showed that even with a change in voice, newly learned words interfered with the recognition of similar known words. These findings show that lexical competition effects from newly learned words spread across different talker voices, which suggests that newly learned words can be sufficiently lexicalized, and abstract with respect to talker voice, without consolidation.


Subject(s)
Inhibition, Psychological , Language , Learning/physiology , Memory, Episodic , Speech Perception/physiology , Adult , Humans
14.
Cognition ; 134: 85-99, 2015 Jan.
Article in English | MEDLINE | ID: mdl-25460382

ABSTRACT

It is well known that familiar words inhibit each other during spoken word recognition. However, we do not know how and under what circumstances newly learned words become integrated with the lexicon in order to engage in this competition. Previous work on word learning has highlighted the importance of offline consolidation (Gaskell & Dumay, 2003) and meaning (Leach & Samuel, 2007) to establish this integration. In two experiments we test the necessity of these factors by examining the inhibition between newly learned items and familiar words immediately after learning. Participants learned a set of nonwords without meanings in active (Experiment 1) or passive (Experiment 2) exposure paradigms. After training, participants performed a visual world paradigm task to assess inhibition from these newly learned items. An analysis of participants' fixations suggested that the newly learned words were able to engage in competition with known words without any consolidation.


Subject(s)
Inhibition, Psychological , Recognition, Psychology/physiology , Speech Perception/physiology , Verbal Learning , Adult , Eye Movement Measurements , Humans , Young Adult
SELECTION OF CITATIONS
SEARCH DETAIL
...