Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 21
Filter
Add more filters










Publication year range
1.
Front Hum Neurosci ; 17: 1225976, 2023.
Article in English | MEDLINE | ID: mdl-37706173

ABSTRACT

Introduction: To perceive speech, our brains process information from different sensory modalities. Previous electroencephalography (EEG) research has established that audio-visual information provides an advantage compared to auditory-only information during early auditory processing. In addition, behavioral research showed that auditory speech perception is not only enhanced by visual information but also by tactile information, transmitted by puffs of air arriving at the skin and aligned with speech. The current EEG study aimed to investigate whether the behavioral benefits of bimodal audio-aerotactile and trimodal audio-visual-aerotactile speech presentation are reflected in cortical auditory event-related neurophysiological responses. Methods: To examine the influence of multimodal information on speech perception, 20 listeners conducted a two-alternative forced-choice syllable identification task at three different signal-to-noise levels. Results: Behavioral results showed increased syllable identification accuracy when auditory information was complemented with visual information, but did not show the same effect for the addition of tactile information. Similarly, EEG results showed an amplitude suppression for the auditory N1 and P2 event-related potentials for the audio-visual and audio-visual-aerotactile modalities compared to auditory and audio-aerotactile presentations of the syllable/pa/. No statistically significant difference was present between audio-aerotactile and auditory-only modalities. Discussion: Current findings are consistent with past EEG research showing a visually induced amplitude suppression during early auditory processing. In addition, the significant neurophysiological effect of audio-visual but not audio-aerotactile presentation is in line with the large benefit of visual information but comparatively much smaller effect of aerotactile information on auditory speech perception previously identified in behavioral research.

2.
Sci Rep ; 12(1): 837, 2022 01 17.
Article in English | MEDLINE | ID: mdl-35039580

ABSTRACT

Face masks slow exhaled air flow and sequester exhaled particles. There are many types of face masks on the market today, each having widely varying fits, filtering, and air redirection characteristics. While particle filtration and flow resistance from masks has been well studied, their effects on speech air flow has not. We built a schlieren system and recorded speech air flow with 14 different face masks, comparing it to mask-less speech. All of the face masks reduced air flow from speech, but some allowed air flow features to reach further than 40 cm from a speaker's lips and nose within a few seconds, and all the face masks allowed some air to escape above the nose. Evidence from available literature shows that distancing and ventilation in higher-risk indoor environment provide more benefit than wearing a face mask. Our own research shows all the masks we tested provide some additional benefit of restricting air flow from a speaker. However, well-fitted mask specifically designed for the purpose of preventing the spread of disease reduce air flow the most. Future research will study the effects of face masks on speech communication in order to facilitate cost/benefit analysis of mask usage in various environments.


Subject(s)
Exhalation/physiology , Filtration/methods , Masks , Speech/physiology , Adult , Disease Transmission, Infectious/prevention & control , Equipment Design , Humans , Male , Young Adult
3.
Sci Rep ; 11(1): 16565, 2021 08 16.
Article in English | MEDLINE | ID: mdl-34400732

ABSTRACT

During locomotion, humans switch gaits from walking to running, and horses from walking to trotting to cantering to galloping, as they increase their movement rate. It is unknown whether gait change leading to a wider movement rate range is limited to locomotive-type behaviours, or instead is a general property of any rate-varying motor system. The tongue during speech provides a motor system that can address this gap. In controlled speech experiments, using phrases containing complex tongue-movement sequences, we demonstrate distinct gaits in tongue movement at different speech rates. As speakers widen their tongue-front displacement range, they gain access to wider speech-rate ranges. At the widest displacement ranges, speakers also produce categorically different patterns for their slowest and fastest speech. Speakers with the narrowest tongue-front displacement ranges show one stable speech-gait pattern, and speakers with widest ranges show two. Critical fluctuation analysis of tongue motion over the time-course of speech revealed these speakers used greater effort at the beginning of phrases-such end-state-comfort effects indicate speech planning. Based on these findings, we expect that categorical motion solutions may emerge in any motor system, providing that system with access to wider movement-rate ranges.


Subject(s)
Speech/physiology , Tongue/physiology , Female , Humans , Male , Movement/physiology , Speech Articulation Tests/instrumentation , Speech Articulation Tests/methods , Time Factors
4.
J Acoust Soc Am ; 146(5): 3495, 2019 11.
Article in English | MEDLINE | ID: mdl-31795693

ABSTRACT

Speech perception is a multi-sensory experience. Visual information enhances [Sumby and Pollack (1954). J. Acoust. Soc. Am. 25, 212-215] and interferes [McGurk and MacDonald (1976). Nature 264, 746-748] with speech perception. Similarly, tactile information, transmitted by puffs of air arriving at the skin and aligned with speech audio, alters [Gick and Derrick (2009). Nature 462, 502-504] auditory speech perception in noise. It has also been shown that aero-tactile information influences visual speech perception when an auditory signal is absent [Derrick, Bicevskis, and Gick (2019a). Front. Commun. Lang. Sci. 3(61), 1-11]. However, researchers have not yet identified the combined influence of aero-tactile, visual, and auditory information on speech perception. The effects of matching and mismatching visual and tactile speech on two-way forced-choice auditory syllable-in-noise classification tasks were tested. The results showed that both visual and tactile information altered the signal-to-noise threshold for accurate identification of auditory signals. Similar to previous studies, the visual component has a strong influence on auditory syllable-in-noise identification, as evidenced by a 28.04 dB improvement in SNR between matching and mismatching visual stimulus presentations. In comparison, the tactile component had a small influence resulting in a 1.58 dB SNR match-mismatch range. The effects of both the audio and tactile information were shown to be additive.

5.
Front Psychol ; 10: 2597, 2019.
Article in English | MEDLINE | ID: mdl-31827453

ABSTRACT

This paper presents the findings of an ultrasound study of 10 New Zealand English and 10 Tongan-speaking trombone players, to determine whether there is an influence of native language speech production on trombone performance. Trombone players' midsagittal tongue shapes were recorded while reading wordlists and during sustained note productions, and tongue surface contours traced. After normalizing to account for differences in vocal tract shape and ultrasound transducer orientation, we used generalized additive mixed models (GAMMs) to estimate average tongue surface shapes used by the players from the two language groups when producing notes at different pitches and intensities, and during the production of the monophthongs in their native languages. The average midsagittal tongue contours predicted by our models show a statistically robust difference at the back of the tongue distinguishing the two groups, where the New Zealand English players display an overall more retracted tongue position; however, tongue shape during playing does not directly map onto vowel tongue shapes as prescribed by the pedagogical literature. While the New Zealand English-speaking participants employed a playing tongue shape approximating schwa and the vowel used in the word 'lot,' the Tongan participants used a tongue shape loosely patterning with the back vowels /o/ and /u/. We argue that these findings represent evidence for native language influence on brass instrument performance; however, this influence seems to be secondary to more basic constraints of brass playing related to airflow requirements and acoustical considerations, with the vocal tract configurations observed across both groups satisfying these conditions in different ways. Our findings furthermore provide evidence for the functional independence of various sections of the tongue and indicate that speech production, itself an acquired motor skill, can influence another skilled behavior via motor memory of vocal tract gestures forming the basis of local optimization processes to arrive at a suitable tongue shape for sustained note production.

6.
J Acoust Soc Am ; 146(3): 1605, 2019 09.
Article in English | MEDLINE | ID: mdl-31590504

ABSTRACT

Integration of auditory and aero-tactile information during speech perception has been documented during two-way closed-choice syllable classification tasks [Gick and Derrick (2009). Nature 462, 502-504], but not during an open-choice task using continuous speech perception [Derrick, O'Beirne, Gorden, De Rybel, Fiasson, and Hay (2016). J. Acoust. Soc. Am. 140(4), 3225]. This study was designed to compare audio-tactile integration during open-choice perception of individual syllables. In addition, this study aimed to compare the effects of place and manner of articulation. Thirty-four untrained participants identified syllables in both auditory-only and audio-tactile conditions in an open-choice paradigm. In addition, forty participants performed a closed-choice perception experiment to allow direct comparison between these two response-type paradigms. Adaptive staircases, as noted by Watson [(1983). Percept. Psychophys. 33(2), 113-120] were used to identify the signal-to-noise ratio for identification accuracy thresholds. The results showed no significant effect of air flow on syllable identification accuracy during the open-choice task, but found a bias towards voiceless identification of labials, and towards voiced identification of velars. Comparison of the open-choice results to those of the closed-choice task show a significant difference between both response types, with audio-tactile integration shown in the closed-choice task, but not in the open-choice task. These results suggest that aero-tactile enhancement of speech perception is dependent on response type demands.


Subject(s)
Phonetics , Speech Perception , Touch Perception , Adolescent , Adult , Air , Female , Humans , Male , Voice/physiology
7.
Annu Rev Linguist ; 5(1): 49-66, 2019 Jan.
Article in English | MEDLINE | ID: mdl-34307767

ABSTRACT

Speech research during recent years has moved progressively away from its traditional focus on audition toward a more multisensory approach. In addition to audition and vision, many somatosenses including proprioception, pressure, vibration and aerotactile sensation are all highly relevant modalities for experiencing and/or conveying speech. In this article, we review both long-standing cross-modal effects stemming from decades of audiovisual speech research as well as new findings related to somatosensory effects. Cross-modal effects in speech perception to date are found to be constrained by temporal congruence and signal relevance, but appear to be unconstrained by spatial congruence. Far from taking place in a one-, two- or even three-dimensional space, the literature reveals that speech occupies a highly multidimensional sensory space. We argue that future research in cross-modal effects should expand to consider each of these modalities both separately and in combination with other modalities in speech.

8.
Article in English | MEDLINE | ID: mdl-35106291

ABSTRACT

Multisensory information is integrated asymmetrically in speech perception: An audio signal can follow video by 240ms, but can precede video by only 60ms, without disrupting the sense of synchronicity (Munhall et al., 1996). Similarly, air flow can follow either audio (Gick et al., 2010) or video (Bicevskis et al., 2016) by a much larger margin than it can precede either while remaining perceptually synchronous. These asymmetric windows of integration have been attributed to the physical properties of the signals; light travels faster than sound (Munhall et al., 1996), and sound travels faster than air flow (Gick et al., 2010). Perceptual windows of integration narrow during development (Hillock-Dunn and Wallace, 2012), but remain wider among people with autism (Wallace and Stevenson, 2014). Here we show that, even among neurotypical adult perceivers, visual-tactile windows of integration are wider and flatter the higher the participant's Autism Quotient (AQ) (Baron-Cohen et al., 2001), a self-report measure of autistic traits. As "pa" is produced with a tiny burst of aspiration (Derrick et al., 2009), we applied light and inaudible air puffs to participants' necks while they watched silent videos of a person saying "ba" or "pa," with puffs presented both synchronously and at varying degrees of asynchrony relative to the recorded plosive release burst, which itself is time-aligned to visible lip opening. All syllables seen along with cutaneous air puffs were more likely to be perceived as "pa." Syllables were perceived as "pa" most often when the air puff occurred 50-100ms after lip opening, with decaying probability as asynchrony increased. Integration was less dependent on time-alignment the higher the participant's AQ. Perceivers integrate event-relevant tactile information in visual speech perception with greater reliance upon event-related accuracy the more they self-describe as neurotypical, supporting the Happé and Frith (2006) weak coherence account of autism spectrum disorder (ASD).

9.
J Acoust Soc Am ; 144(5): EL392, 2018 11.
Article in English | MEDLINE | ID: mdl-30522328

ABSTRACT

When using ultrasound imaging of the tongue for speech recording/research, submental transducer stabilization is required to prevent the ultrasound transducer from translating or rotating in relation to the tongue. An iterative prototype of a lightweight three-dimensional-printable wearable ultrasound transducer stabilization system that allows flexible jaw motion and free head movement is presented. The system is completely non-metallic, eliminating interference with co-recorded signals, thus permitting co-collection and co-registration with articulometry systems. A motion study of the final version demonstrates that transducer rotation is limited to 1.25° and translation to 2.5 mm-well within accepted tolerances.


Subject(s)
Printing, Three-Dimensional/instrumentation , Speech/physiology , Tongue/diagnostic imaging , Ultrasonography/methods , Germany/ethnology , Head Movements/physiology , Humans , Jaw/physiology , Male , Speech Articulation Tests/methods , Transducers
10.
J Speech Lang Hear Res ; 59(6): S1566-S1574, 2016 12 01.
Article in English | MEDLINE | ID: mdl-28002837

ABSTRACT

Purpose: Models of speech production often abstract away from shared physiology in pitch control and lingual articulation, positing independent control of tone and vowel units. We assess the validity of this assumption in Mandarin Chinese by evaluating the stability of lingual articulation for vowels across variation in tone. Method: Electromagnetic articulography was used to track flesh points on the tongue (tip, body, dorsum), lips, and jaw while native Mandarin speakers (n = 6) produced 3 vowels, /pa/, /pi/, /pu/, combined with 4 Mandarin tones: T1 "high," T2 "rising," T3 "low," and T4 "falling." Results: Consistent with physiological expectations, tones that begin low, T2 and T3, conditioned a lower position of the tongue body for the vowel /a/. For the vowel /i/, we found the opposite effect, whereby tones that begin low, T2 and T3, conditioned a higher tongue body position. Conclusions: The physiology of pitch control exerts systematic variation on the lingual articulation of /a/ across tones. The effects of tone on /i/ articulation are in the opposite direction predicted by physiological considerations. Physiologically arbitrary variation of the type observed for /i/ challenges the assumption that phonetic patterns can be determined by independent control of tone (source) and vowel (filter) production units.


Subject(s)
Jaw , Motor Activity , Phonetics , Speech Acoustics , Tongue , Adult , Biomechanical Phenomena , Female , Humans , Jaw/physiology , Male , Motor Activity/physiology , Motor Skills/physiology , Tongue/physiology , Young Adult
11.
J Acoust Soc Am ; 140(5): 3531, 2016 Nov.
Article in English | MEDLINE | ID: mdl-27908052

ABSTRACT

Audio-visual [McGurk and MacDonald (1976). Nature 264, 746-748] and audio-tactile [Gick and Derrick (2009). Nature 462(7272), 502-504] speech stimuli enhance speech perception over audio stimuli alone. In addition, multimodal speech stimuli form an asymmetric window of integration that is consistent with the relative speeds of the various signals [Munhall, Gribble, Sacco, and Ward (1996). Percept. Psychophys. 58(3), 351-362; Gick, Ikegami, and Derrick (2010). J. Acoust. Soc. Am. 128(5), EL342-EL346]. In this experiment, participants were presented video of faces producing /pa/ and /ba/ syllables, both alone and with air puffs occurring synchronously and at different timings up to 300 ms before and after the stop release. Perceivers were asked to identify the syllable they perceived, and were more likely to respond that they perceived /pa/ when air puffs were present, with asymmetrical preference for puffs following the video signal-consistent with the relative speeds of visual and air puff signals. The results demonstrate that visual-tactile integration of speech perception occurs much as it does with audio-visual and audio-tactile stimuli. This finding contributes to the understanding of multimodal speech perception, lending support to the idea that speech is not perceived as an audio signal that is supplemented by information from other modes, but rather that primitives of speech perception are, in principle, modality neutral.


Subject(s)
Speech Perception , Acoustic Stimulation , Adolescent , Adult , Female , Humans , Male , Speech , Touch , Visual Perception , Young Adult
12.
Eur J Orthod ; 38(1): 90-95, 2016 Feb.
Article in English | MEDLINE | ID: mdl-25888532

ABSTRACT

BACKGROUND: First, first permanent molars (FPMs) of poor prognosis are often planned for extraction at an 'ideal time' so that second permanent molars (SPMs) erupt favourably to replace them. However for lower FPM extractions, timing is not an accurate predictor of success. OBJECTIVE: The aim of this study was to identify additional radiographic factors that could better predict the degree of spontaneous space closure of the lower SPM following FPM extraction. METHODS: Data from a previous study of 127 lower SPMs from 66 patients was re-analysed by incorporating additional radiographic factors. These included calcification stage of the bifurcation of the SPM, position of the second premolar, mesial angulation of SPM in relation to the FPM, and presence of the third permanent molar. Results were analysed using ordered logistic regression. RESULTS: Only 58 per cent of FPMs extracted at the 'ideal time' (SPM development at Demirjian stage E) had complete space closure. The best outcomes resulted from a combination of SPMs not at Demirjian development stage G, together with the presence of mesial angulation of the SPM and presence of the third permanent molar, where 85 per cent of those cases had complete space closure. CONCLUSIONS: Apart from extraction timing of the FPM, consideration must also be given to the presence of the third permanent molar and angulation of the SPM in order to ensure a reliable degree of spontaneous space closure of the lower SPM.


Subject(s)
Molar/surgery , Tooth Extraction/methods , Tooth Migration/physiopathology , Adolescent , Bicuspid/diagnostic imaging , Bicuspid/pathology , Child , Female , Humans , Male , Molar/diagnostic imaging , Molar/pathology , Molar, Third/diagnostic imaging , Molar, Third/pathology , Orthodontics, Corrective/methods , Prognosis , Radiography, Dental/methods , Retrospective Studies , Tooth Eruption/physiology , Tooth Migration/diagnostic imaging , Tooth Migration/pathology
13.
J Acoust Soc Am ; 138(2): EL161-6, 2015 Aug.
Article in English | MEDLINE | ID: mdl-26328743

ABSTRACT

This study proposes a method of superimposing a physical palatal profile, extracted from a speaker's maxillary impression, onto real-time mid-sagittal articulatory data. A palatal/dental profile is first obtained by three-dimensional-scanning the maxillary impression of the speaker. Then a high resolution mid-sagittal palatal line, extracted from the profile, is sub-divided into articulatory zones and superimposed, by Iterative Closest Point algorithm, onto reconstructed palatal traces in electromagnetic articulometric (EMA) data. Evaluations were carried out by comparing consonant targets elicited by EMA with the proposed method and by static palatography. The proposed method yields accurate results, as supported by palatography.


Subject(s)
Image Processing, Computer-Assisted/methods , Imaging, Three-Dimensional , Maxilla/anatomy & histology , Models, Dental , Palate/anatomy & histology , Phonetics , Speech Articulation Tests/methods , Adult , Alginates , Algorithms , Analog-Digital Conversion , Calcium Sulfate , Computer Systems , Dental Impression Materials , Dental Impression Technique , Female , Fiducial Markers , Glucuronic Acid , Hexuronic Acids , Humans , Male , Maxilla/diagnostic imaging , Palate/diagnostic imaging , Radiography , Speech Articulation Tests/instrumentation
14.
J Acoust Soc Am ; 137(3): 1493-502, 2015 Mar.
Article in English | MEDLINE | ID: mdl-25786960

ABSTRACT

The assumption that units of speech production bear a one-to-one relationship to speech motor actions pervades otherwise widely varying theories of speech motor behavior. This speech production and simulation study demonstrates that commonly occurring flap sequences may violate this assumption. In the word "Saturday," a sequence of three sounds may be produced using a single, cyclic motor action. Under this view, the initial upward tongue tip motion, starting with the first vowel and moving to contact the hard palate on the way to a retroflex position, is under active muscular control, while the downward movement of the tongue tip, including the second contact with the hard palate, results from gravity and elasticity during tongue muscle relaxation. This sequence is reproduced using a three-dimensional computer simulation of human vocal tract biomechanics and differs greatly from other observed sequences for the same word, which employ multiple targeted speech motor actions. This outcome suggests that a goal of a speaker is to produce an entire sequence in a biomechanically efficient way at the expense of maintaining parity within the individual parts of the sequence.


Subject(s)
Larynx/physiology , Motor Activity , Phonetics , Speech Acoustics , Tongue/physiology , Voice Quality , Acoustics , Adolescent , Adult , Biomechanical Phenomena , Computer Simulation , Elasticity , Female , Finite Element Analysis , Humans , Imaging, Three-Dimensional , Male , Models, Biological , Movement , Muscle Contraction , Nonlinear Dynamics , Palate, Hard/physiology , Speech Production Measurement , Time Factors , Tongue/diagnostic imaging , Ultrasonography , Young Adult
15.
J Acoust Soc Am ; 138(6): EL509-14, 2015 Dec.
Article in English | MEDLINE | ID: mdl-26723359

ABSTRACT

Tongue surface measurements from midsagittal ultrasound scans are effectively arcs with deviations representing tongue shape, but smoothing-spline analysis of variances (SSANOVAs) assume variance around a horizontal line. Therefore, calculating SSANOVA average curves of tongue traces in Cartesian Coordinates [Davidson, J. Acoust. Soc. Am. 120(1), 407-415 (2006)] creates errors that are compounded at tongue tip and root where average tongue shape deviates most from a horizontal line. This paper introduces a method for transforming data into polar coordinates similar to the technique by Mielke [J. Acoust. Soc. Am. 137(5), 2858-2869 (2015)], but using the virtual origin of a radial ultrasound transducer as the polar origin-allowing data conversion in a manner that is robust against between-subject and between-session variability.

16.
Phonetica ; 71(3): 183-200, 2014.
Article in English | MEDLINE | ID: mdl-25790787

ABSTRACT

Applying the 'end-state comfort' hypothesis of Rosenbaum et al. [J Exp Psych Learn Mem Cogn 1992;18:1058; Acta Psychol (Amst) 1996;94:59] to tongue motion provides evidence of long-distance subphonemic planning in speech. Speakers' tongue postures may anticipate upcoming speech up to three segments, two syllables, and a morpheme or word boundary later. We used M-mode ultrasound imaging to measure the direction of tongue tip/blade movements for known variants of flap/tap allophones of North American English /t/ and /d/. Results show that speakers produce different flap variants early in words or word sequences so as to facilitate the kinematic needs of flap/tap or other /r/ variants that appear later in the word or word sequence. Similar results were also observed across word boundaries, indicating that this is not a lexical effect.


Subject(s)
Language , Phonetics , Speech , Tongue/physiology , Humans , North America
17.
Multisens Res ; 26(5): 405-16, 2013.
Article in English | MEDLINE | ID: mdl-24649526

ABSTRACT

Tactile sensations at extreme distal body locations can integrate with auditory information to alter speech perception among uninformed and untrained listeners. Inaudible air puffs were applied to participants' ankles, simultaneously with audible syllables having aspirated and unaspirated stop onsets. Syllables heard simultaneously with air puffs were more likely to be heard as aspirated. These results demonstrate that event-appropriate information from distal parts of the body integrates in speech perception, even without frequent or robust location-specific experience. In addition, overall performance was significantly better for those with hair on their ankles, which suggests that the presence of hair may help establish signal relevance, and so aid in multi-modal speech perception.


Subject(s)
Auditory Perception/physiology , Touch/physiology , Acoustic Stimulation/methods , Adult , Female , Humans , Male , Skin , Young Adult
18.
J Acoust Soc Am ; 131(5): EL355-60, 2012 May.
Article in English | MEDLINE | ID: mdl-22559452

ABSTRACT

This study reports an investigation of the well-known context-dependent variation in English /r/ using a biomechanical tongue-jaw-hyoid model. The simulation results show that preferred /r/ variants require less volume displacement, relative strain, and relative muscle stress than variants that are not preferred. This study also uncovers a previously unknown mechanism in tongue biomechanics for /r/ production: Torque in the sagittal plane about the mental spine. This torque enables raising of the tongue anterior for retroflexed [Symbol: see text] by activation of hyoglossus and relaxation of anterior genioglossus. The results provide a deeper understanding of the articulatory factors that govern contextual phonetic variation.


Subject(s)
Jaw/physiology , Phonetics , Tongue/physiology , Biomechanical Phenomena , Humans , Models, Biological , Movement/physiology , Muscle, Skeletal/physiology , Stress, Physiological/physiology
19.
J Acoust Soc Am ; 128(5): EL342-6, 2010 Nov.
Article in English | MEDLINE | ID: mdl-21110549

ABSTRACT

Asynchronous cross-modal information is integrated asymmetrically in audio-visual perception. To test whether this asymmetry generalizes across modalities, auditory (aspirated "pa" and unaspirated "ba" stops) and tactile (slight, inaudible, cutaneous air puffs) signals were presented synchronously and asynchronously. Results were similar to previous AV studies: the temporal window of integration for the enhancement effect (but not the interference effect) was asymmetrical, allowing up to 200 ms of asynchrony when the puff followed the audio signal, but only up to 50 ms when the puff preceded the audio signal. These findings suggest that perceivers accommodate differences in physical transmission speed of different multimodal signals.


Subject(s)
Auditory Threshold/physiology , Peripheral Nerves/physiology , Speech Perception/physiology , Touch/physiology , Acoustic Stimulation , Adult , Air Movements , Cues , Humans , Physical Stimulation
20.
Nature ; 462(7272): 502-4, 2009 Nov 26.
Article in English | MEDLINE | ID: mdl-19940925

ABSTRACT

Visual information from a speaker's face can enhance or interfere with accurate auditory perception. This integration of information across auditory and visual streams has been observed in functional imaging studies, and has typically been attributed to the frequency and robustness with which perceivers jointly encounter event-specific information from these two modalities. Adding the tactile modality has long been considered a crucial next step in understanding multisensory integration. However, previous studies have found an influence of tactile input on speech perception only under limited circumstances, either where perceivers were aware of the task or where they had received training to establish a cross-modal mapping. Here we show that perceivers integrate naturalistic tactile information during auditory speech perception without previous training. Drawing on the observation that some speech sounds produce tiny bursts of aspiration (such as English 'p'), we applied slight, inaudible air puffs on participants' skin at one of two locations: the right hand or the neck. Syllables heard simultaneously with cutaneous air puffs were more likely to be heard as aspirated (for example, causing participants to mishear 'b' as 'p'). These results demonstrate that perceivers integrate event-relevant tactile information in auditory perception in much the same way as they do visual information.


Subject(s)
Air Pressure , Sound , Speech Perception/physiology , Touch/physiology , Acoustic Stimulation , Cues , Hair/physiology , Hand/physiology , Humans , Neck/physiology , Research Design , Skin Physiological Phenomena
SELECTION OF CITATIONS
SEARCH DETAIL
...