Electrophysiological indicators of phonetic and non-phonetic multisensory interactions during audiovisual speech perception

被引:145
作者
Klucharev, V [1 ]
Möttönen, R [1 ]
Sams, M [1 ]
机构
[1] Aalto Univ, Lab Computat Engn, FIN-02015 Espoo, Finland
来源
COGNITIVE BRAIN RESEARCH | 2003年 / 18卷 / 01期
基金
芬兰科学院;
关键词
audiovisual interaction; multisensory processing; event-related potential (ERP); speech perception;
D O I
10.1016/j.cogbrainres.2003.09.004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We studied the interactions in neural processing of auditory and visual speech by recording event-related brain potentials (ERPs). Unisensory (auditory-A and visual-V) and audiovisual (AV) vowels were presented to 11 subjects. AV vowels were phonetically either congruent (e.g., acoustic /a/ and visual /a/) or incongruent (e.g., acoustic /a/ and visual /y/. ERPs to AV stimuli and the sum of the ERPs to A and V stimuli (A + V) were compared. Similar ERPs to AV and A + V were hypothesized to indicate independent processing of A and V stimuli. Differences on the other hand would suggest AV interactions. Three deflections, the first peaking at about 85 ms after the A stimulus onset, were significantly larger in the ERPs to A + V than in the ERPs to both congruent and incongruent AV stimuli. We suggest that these differences reflect AV interactions in the processing of general, non-phonetic, features shared by the acoustic and visual stimulus (spatial location, coincidence in time). The first difference in the ERPs to incongruent and congruent AV vowels peaked at 155 ms from the A stimuli onset. This and two later differences are suggested to reflect interactions at phonetic level. The early general AV interactions probably reflect modified activity in the sensory-specific cortices, whereas the later phonetic AV interactions are likely generated in the heteromodal cortices. Thus, our results suggest that sensory-specific and heteromodal brain regions participate in AV speech integration at separate latencies and are sensitive to different features of A and V speech stimuli. (C) 2003 Elsevier B.V. All rights reserved.
引用
收藏
页码:65 / 75
页数:11
相关论文
共 41 条
[1]   EFFECTS OF AN AUDITORY SIGNAL ON VISUAL REACTION TIME [J].
BERNSTEIN, IH ;
CLARK, MH ;
EDELSTEIN, BA .
JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 1969, 80 (3P1) :567-+
[2]   Human temporal lobe activation by speech and nonspeech sounds [J].
Binder, JR ;
Frost, JA ;
Hammeke, TA ;
Bellgowan, PSF ;
Springer, JA ;
Kaufman, JN ;
Possing, ET .
CEREBRAL CORTEX, 2000, 10 (05) :512-528
[3]   Neural correlates of auditory-visual stimulus onset asynchrony detection [J].
Bushara, KO ;
Grafman, J ;
Hallett, M .
JOURNAL OF NEUROSCIENCE, 2001, 21 (01) :300-304
[4]   Evidence from functional magnetic resonance imaging of crossmodal binding in the human heteromodal cortex [J].
Calvert, GA ;
Campbell, R ;
Brammer, MJ .
CURRENT BIOLOGY, 2000, 10 (11) :649-657
[5]   Response amplification in sensory-specific cortices during crossmodal binding [J].
Calvert, GA ;
Brammer, MJ ;
Bullmore, ET ;
Campbell, R ;
Iversen, SD ;
David, AS .
NEUROREPORT, 1999, 10 (12) :2619-2623
[6]   Reading speech from still and moving faces: The neural substrates of visible speech [J].
Calvert, GA ;
Campbell, R .
JOURNAL OF COGNITIVE NEUROSCIENCE, 2003, 15 (01) :57-70
[7]   Detection of audio-visual integration sites in humans by application of electrophysiological criteria to the BOLD effect [J].
Calvert, GA ;
Hansen, PC ;
Iversen, SD ;
Brammer, MJ .
NEUROIMAGE, 2001, 14 (02) :427-438
[8]   Crossmodal processing in the human brain: Insights from functional neuroimaging studies [J].
Calvert, GA .
CEREBRAL CORTEX, 2001, 11 (12) :1110-1123
[9]   Cortical substrates for the perception of face actions: an fMRI study of the specificity of activation for seen speech and for meaningless lower-face acts (gurning) [J].
Campbell, R ;
MacSweeney, M ;
Surguladze, S ;
Calvert, G ;
McGuire, P ;
Suckling, J ;
Brammer, MJ ;
David, AS .
COGNITIVE BRAIN RESEARCH, 2001, 12 (02) :233-243
[10]   Mismatch negativity evoked by the McGurk-MacDonald effect: a phonetic representation within short-term memory [J].
Colin, C ;
Radeau, M ;
Soquet, A ;
Demolin, D ;
Colin, F ;
Deltenre, P .
CLINICAL NEUROPHYSIOLOGY, 2002, 113 (04) :495-506