1
|
Modulation of early auditory processing by visual information: Prediction or bimodal integration? Atten Percept Psychophys 2021; 83:1538-1551. [PMID: 33506354 PMCID: PMC8084811 DOI: 10.3758/s13414-021-02240-1] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/29/2020] [Indexed: 11/08/2022]
Abstract
What happens if a visual cue misleads auditory expectations? Previous studies revealed an early visuo-auditory incongruency effect, so-called incongruency response (IR) of the auditory event-related brain potential (ERP), occurring 100 ms after onset of the sound being incongruent to the preceding visual cue. So far, this effect has been ascribed to reflect the mismatch between auditory sensory expectation activated by visual predictive information and the actual sensory input. Thus, an IR should be confined to an asynchronous presentation of visual cue and sound. Alternatively, one could argue that frequently presented congruent visual-cue-sound combinations are integrated into a bimodal representation whereby violation of the visual-auditory relationship results in a bimodal feature mismatch (the IR should be obtained with asynchronous and with synchronous presentation). In an asynchronous condition, an either high-pitched or low-pitched sound was preceded by a visual note symbol presented above or below a fixation cross (90% congruent; 10% incongruent), while in a synchronous condition, both were presented simultaneously. High-pitched and low-pitched sounds were presented with different probabilities (83% vs. 17%) to form a strong association between bimodal stimuli. In both conditions, tones with pitch incongruent with the location of the note symbols elicited incongruency effects in the N2 and P3 ERPs; however, the IR was only elicited in the asynchronous condition. This finding supports the sensorial prediction error hypothesis stating that the amplitude of the auditory ERP 100 ms after sound onset is enhanced in response to unexpected compared with expected but otherwise identical sounds.
Collapse
|
2
|
Sorati M, Behne DM. Considerations in Audio-Visual Interaction Models: An ERP Study of Music Perception by Musicians and Non-musicians. Front Psychol 2021; 11:594434. [PMID: 33551911 PMCID: PMC7854916 DOI: 10.3389/fpsyg.2020.594434] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2020] [Accepted: 12/03/2020] [Indexed: 11/13/2022] Open
Abstract
Previous research with speech and non-speech stimuli suggested that in audiovisual perception, visual information starting prior to the onset of corresponding sound can provide visual cues, and form a prediction about the upcoming auditory sound. This prediction leads to audiovisual (AV) interaction. Auditory and visual perception interact and induce suppression and speeding up of the early auditory event-related potentials (ERPs) such as N1 and P2. To investigate AV interaction, previous research examined N1 and P2 amplitudes and latencies in response to audio only (AO), video only (VO), audiovisual, and control (CO) stimuli, and compared AV with auditory perception based on four AV interaction models (AV vs. AO+VO, AV-VO vs. AO, AV-VO vs. AO-CO, AV vs. AO). The current study addresses how different models of AV interaction express N1 and P2 suppression in music perception. Furthermore, the current study took one step further and examined whether previous musical experience, which can potentially lead to higher N1 and P2 amplitudes in auditory perception, influenced AV interaction in different models. Musicians and non-musicians were presented the recordings (AO, AV, VO) of a keyboard /C4/ key being played, as well as CO stimuli. Results showed that AV interaction models differ in their expression of N1 and P2 amplitude and latency suppression. The calculation of model (AV-VO vs. AO) and (AV-VO vs. AO-CO) has consequences for the resulting N1 and P2 difference waves. Furthermore, while musicians, compared to non-musicians, showed higher N1 amplitude in auditory perception, suppression of amplitudes and latencies for N1 and P2 was similar for the two groups across the AV models. Collectively, these results suggest that when visual cues from finger and hand movements predict the upcoming sound in AV music perception, suppression of early ERPs is similar for musicians and non-musicians. Notably, the calculation differences across models do not lead to the same pattern of results for N1 and P2, demonstrating that the four models are not interchangeable and are not directly comparable.
Collapse
Affiliation(s)
- Marzieh Sorati
- Department of Psychology, Norwegian University of Science and Technology, Trondheim, Norway
| | - Dawn M Behne
- Department of Psychology, Norwegian University of Science and Technology, Trondheim, Norway
| |
Collapse
|
3
|
Kobayashi Y. Effect of Cotton Roll Biting on Auditory Evoked Magnetic Fields. THE BULLETIN OF TOKYO DENTAL COLLEGE 2017; 58:137-143. [DOI: 10.2209/tdcpublication.2016-2000] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
4
|
Nakamura I, Hirano Y, Ohara N, Hirano S, Ueno T, Tsuchimoto R, Kanba S, Onitsuka T. Early integration processing between faces and vowel sounds in human brain: an MEG investigation. Neuropsychobiology 2016; 71:187-95. [PMID: 26044647 DOI: 10.1159/000377680] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/02/2014] [Accepted: 02/03/2015] [Indexed: 11/19/2022]
Abstract
OBJECTIVE Unconscious fast integration of face and voice information is a crucial brain function necessary for communicating effectively with others. Here, we investigated for evidence of rapid face-voice integration in the auditory cortex. METHODS Magnetic fields (P50m and N100m) evoked by visual stimuli (V), auditory stimuli (A) and audiovisual stimuli (VA), i.e. by face, vowel and simultaneous vowel-face stimuli, were recorded in 22 healthy subjects. Magnetoencephalographic data from 28 channels around bilateral auditory cortices were analyzed. RESULTS In both hemispheres, AV - V showed significantly larger P50m amplitudes than A. Additionally, compared with A, the N100m amplitudes and dipole moments of AV - V were significantly smaller in the left hemisphere, but not in the right hemisphere. CONCLUSIONS Differential changes in P50m (bilateral) and N100m (left hemisphere) that occur when V (faces) are associated with A (vowel sounds) indicate that AV (face-voice) integration occurs in early processing, likely enabling us to communicate effectively in our lives.
Collapse
Affiliation(s)
- Itta Nakamura
- Department of Neuropsychiatry, Graduate School of Medical Sciences, Kyushu University, Fukuoka, Japan
| | | | | | | | | | | | | | | |
Collapse
|
5
|
Alsius A, Möttönen R, Sams ME, Soto-Faraco S, Tiippana K. Effect of attentional load on audiovisual speech perception: evidence from ERPs. Front Psychol 2014; 5:727. [PMID: 25076922 PMCID: PMC4097954 DOI: 10.3389/fpsyg.2014.00727] [Citation(s) in RCA: 49] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2014] [Accepted: 06/23/2014] [Indexed: 11/13/2022] Open
Abstract
Seeing articulatory movements influences perception of auditory speech. This is often reflected in a shortened latency of auditory event-related potentials (ERPs) generated in the auditory cortex. The present study addressed whether this early neural correlate of audiovisual interaction is modulated by attention. We recorded ERPs in 15 subjects while they were presented with auditory, visual, and audiovisual spoken syllables. Audiovisual stimuli consisted of incongruent auditory and visual components known to elicit a McGurk effect, i.e., a visually driven alteration in the auditory speech percept. In a Dual task condition, participants were asked to identify spoken syllables whilst monitoring a rapid visual stream of pictures for targets, i.e., they had to divide their attention. In a Single task condition, participants identified the syllables without any other tasks, i.e., they were asked to ignore the pictures and focus their attention fully on the spoken syllables. The McGurk effect was weaker in the Dual task than in the Single task condition, indicating an effect of attentional load on audiovisual speech perception. Early auditory ERP components, N1 and P2, peaked earlier to audiovisual stimuli than to auditory stimuli when attention was fully focused on syllables, indicating neurophysiological audiovisual interaction. This latency decrement was reduced when attention was loaded, suggesting that attention influences early neural processing of audiovisual speech. We conclude that reduced attention weakens the interaction between vision and audition in speech.
Collapse
Affiliation(s)
- Agnès Alsius
- Psychology Department, Queen's University Kingston, ON, Canada
| | - Riikka Möttönen
- Department of Experimental Psychology, University of Oxford Oxford, UK
| | - Mikko E Sams
- Brain and Mind Laboratory, School of Science, Aalto University Espoo, Finland
| | - Salvador Soto-Faraco
- Institut Català de Recerca i Estudis Avançats Barcelona, Spain ; Brain and Cognition Center, Universitat Pompeu Fabra Barcelona, Spain
| | - Kaisa Tiippana
- Institute of Behavioural Sciences, University of Helsinki Helsinki, Finland
| |
Collapse
|
6
|
Paris T, Kim J, Davis C. Visual speech form influences the speed of auditory speech processing. BRAIN AND LANGUAGE 2013; 126:350-356. [PMID: 23942046 DOI: 10.1016/j.bandl.2013.06.008] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/05/2013] [Revised: 05/28/2013] [Accepted: 06/29/2013] [Indexed: 06/02/2023]
Abstract
An important property of visual speech (movements of the lips and mouth) is that it generally begins before auditory speech. Research using brain-based paradigms has demonstrated that seeing visual speech speeds up the activation of the listener's auditory cortex but it is not clear whether these observed neural processes link to behaviour. It was hypothesized that the very early portion of visual speech (occurring before auditory speech) will allow listeners to predict the following auditory event and so facilitate the speed of speech perception. This was tested in the current behavioural experiments. Further, we tested whether the salience of the visual speech played a role in this speech facilitation effect (Experiment 1). We also determined the relative contributions that visual form (what) and temporal (when) cues made (Experiment 2). The results showed that visual speech cues facilitated response times and that this was based on form rather than temporal cues.
Collapse
Affiliation(s)
- Tim Paris
- The MARCS Institute, University of Western Sydney, Sydney, Australia.
| | | | | |
Collapse
|
7
|
Hertrich I, Dietrich S, Ackermann H. How can audiovisual pathways enhance the temporal resolution of time-compressed speech in blind subjects? Front Psychol 2013; 4:530. [PMID: 23966968 PMCID: PMC3745084 DOI: 10.3389/fpsyg.2013.00530] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2013] [Accepted: 07/26/2013] [Indexed: 11/13/2022] Open
Abstract
In blind people, the visual channel cannot assist face-to-face communication via lipreading or visual prosody. Nevertheless, the visual system may enhance the evaluation of auditory information due to its cross-links to (1) the auditory system, (2) supramodal representations, and (3) frontal action-related areas. Apart from feedback or top-down support of, for example, the processing of spatial or phonological representations, experimental data have shown that the visual system can impact auditory perception at more basic computational stages such as temporal signal resolution. For example, blind as compared to sighted subjects are more resistant against backward masking, and this ability appears to be associated with activity in visual cortex. Regarding the comprehension of continuous speech, blind subjects can learn to use accelerated text-to-speech systems for "reading" texts at ultra-fast speaking rates (>16 syllables/s), exceeding by far the normal range of 6 syllables/s. A functional magnetic resonance imaging study has shown that this ability, among other brain regions, significantly covaries with BOLD responses in bilateral pulvinar, right visual cortex, and left supplementary motor area. Furthermore, magnetoencephalographic measurements revealed a particular component in right occipital cortex phase-locked to the syllable onsets of accelerated speech. In sighted people, the "bottleneck" for understanding time-compressed speech seems related to higher demands for buffering phonological material and is, presumably, linked to frontal brain structures. On the other hand, the neurophysiological correlates of functions overcoming this bottleneck, seem to depend upon early visual cortex activity. The present Hypothesis and Theory paper outlines a model that aims at binding these data together, based on early cross-modal pathways that are already known from various audiovisual experiments on cross-modal adjustments during space, time, and object recognition.
Collapse
Affiliation(s)
- Ingo Hertrich
- Department of General Neurology, Center of Neurology, Hertie Institute for Clinical Brain Research, University of Tübingen Tübingen, Germany
| | | | | |
Collapse
|
8
|
Abstract
AbstractThere is a strong interaction between multisensory processing and the neuroplasticity of the human brain. On one hand, recent research demonstrates that experience and training in various domains modifies how information from the different senses is integrated; and, on the other hand multisensory training paradigms seem to be particularly effective in driving functional and structural plasticity. Multisensory training affects early sensory processing within separate sensory domains, as well as the functional and structural connectivity between uni- and multisensory brain regions. In this review, we discuss the evidence for interactions of multisensory processes and brain plasticity and give an outlook on promising clinical applications and open questions.
Collapse
|
9
|
Maguinness C, Setti A, Burke KE, Kenny RA, Newell FN. The effect of combined sensory and semantic components on audio-visual speech perception in older adults. Front Aging Neurosci 2011; 3:19. [PMID: 22207848 PMCID: PMC3244611 DOI: 10.3389/fnagi.2011.00019] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2011] [Accepted: 11/28/2011] [Indexed: 11/13/2022] Open
Abstract
Previous studies have found that perception in older people benefits from multisensory over unisensory information. As normal speech recognition is affected by both the auditory input and the visual lip movements of the speaker, we investigated the efficiency of audio and visual integration in an older population by manipulating the relative reliability of the auditory and visual information in speech. We also investigated the role of the semantic context of the sentence to assess whether audio–visual integration is affected by top-down semantic processing. We presented participants with audio–visual sentences in which the visual component was either blurred or not blurred. We found that there was a greater cost in recall performance for semantically meaningless speech in the audio–visual ‘blur’ compared to audio–visual ‘no blur’ condition and this effect was specific to the older group. Our findings have implications for understanding how aging affects efficient multisensory integration for the perception of speech and suggests that multisensory inputs may benefit speech perception in older adults when the semantic content of the speech is unpredictable.
Collapse
|
10
|
Lipreading and covert speech production similarly modulate human auditory-cortex responses to pure tones. J Neurosci 2010; 30:1314-21. [PMID: 20107058 DOI: 10.1523/jneurosci.1950-09.2010] [Citation(s) in RCA: 33] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
Watching the lips of a speaker enhances speech perception. At the same time, the 100 ms response to speech sounds is suppressed in the observer's auditory cortex. Here, we used whole-scalp 306-channel magnetoencephalography (MEG) to study whether lipreading modulates human auditory processing already at the level of the most elementary sound features, i.e., pure tones. We further envisioned the temporal dynamics of the suppression to tell whether the effect is driven by top-down influences. Nineteen subjects were presented with 50 ms tones spanning six octaves (125-8000 Hz) (1) during "lipreading," i.e., when they watched video clips of silent articulations of Finnish vowels /a/, /i/, /o/, and /y/, and reacted to vowels presented twice in a row; (2) during a visual control task; (3) during a still-face passive control condition; and (4) in a separate experiment with a subset of nine subjects, during covert production of the same vowels. Auditory-cortex 100 ms responses (N100m) were equally suppressed in the lipreading and covert-speech-production tasks compared with the visual control and baseline tasks; the effects involved all frequencies and were most prominent in the left hemisphere. Responses to tones presented at different times with respect to the onset of the visual articulation showed significantly increased N100m suppression immediately after the articulatory gesture. These findings suggest that the lipreading-related suppression in the auditory cortex is caused by top-down influences, possibly by an efference copy from the speech-production system, generated during both own speech and lipreading.
Collapse
|
11
|
Electrophysiological (EEG, sEEG, MEG) evidence for multiple audiovisual interactions in the human auditory cortex. Hear Res 2009; 258:143-51. [DOI: 10.1016/j.heares.2009.06.016] [Citation(s) in RCA: 65] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/31/2009] [Revised: 06/24/2009] [Accepted: 06/25/2009] [Indexed: 11/23/2022]
|
12
|
Hertrich I, Mathiak K, Lutzenberger W, Ackermann H. Time course of early audiovisual interactions during speech and nonspeech central auditory processing: a magnetoencephalography study. J Cogn Neurosci 2009; 21:259-74. [PMID: 18510440 DOI: 10.1162/jocn.2008.21019] [Citation(s) in RCA: 12] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Cross-modal fusion phenomena suggest specific interactions of auditory and visual sensory information both within the speech and nonspeech domains. Using whole-head magnetoencephalography, this study recorded M50 and M100 fields evoked by ambiguous acoustic stimuli that were visually disambiguated to perceived /ta/ or /pa/ syllables. As in natural speech, visual motion onset preceded the acoustic signal by 150 msec. Control conditions included visual and acoustic nonspeech signals as well as visual-only and acoustic-only stimuli. (a) Both speech and nonspeech motion yielded a consistent attenuation of the auditory M50 field, suggesting a visually induced "preparatory baseline shift" at the level of the auditory cortex. (b) Within the temporal domain of the auditory M100 field, visual speech and nonspeech motion gave rise to different response patterns (nonspeech: M100 attenuation; visual /pa/: left-hemisphere M100 enhancement; /ta/: no effect). (c) These interactions could be further decomposed using a six-dipole model. One of these three pairs of dipoles (V270) was fitted to motion-induced activity at a latency of 270 msec after motion onset, that is, the time domain of the auditory M100 field, and could be attributed to the posterior insula. This dipole source responded to nonspeech motion and visual /pa/, but was found suppressed in the case of visual /ta/. Such a nonlinear interaction might reflect the operation of a binary distinction between the marked phonological feature "labial" versus its underspecified competitor "coronal." Thus, visual processing seems to be shaped by linguistic data structures even prior to its fusion with auditory information channel.
Collapse
|
13
|
Miki K, Kida T, Tanaka E, Nagata O, Kakigi R. The impact of visual movement on auditory cortical responses: a magnetoencephalographic study. Exp Brain Res 2009; 194:597-604. [DOI: 10.1007/s00221-009-1735-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2008] [Accepted: 02/05/2009] [Indexed: 11/30/2022]
|
14
|
Stekelenburg JJ, Vroomen J. Neural correlates of multisensory integration of ecologically valid audiovisual events. J Cogn Neurosci 2008; 19:1964-73. [PMID: 17892381 DOI: 10.1162/jocn.2007.19.12.1964] [Citation(s) in RCA: 244] [Impact Index Per Article: 15.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
A question that has emerged over recent years is whether audiovisual (AV) speech perception is a special case of multi-sensory perception. Electrophysiological (ERP) studies have found that auditory neural activity (N1 component of the ERP) induced by speech is suppressed and speeded up when a speech sound is accompanied by concordant lip movements. In Experiment 1, we show that this AV interaction is not speech-specific. Ecologically valid nonspeech AV events (actions performed by an actor such as handclapping) were associated with a similar speeding-up and suppression of auditory N1 amplitude as AV speech (syllables). Experiment 2 demonstrated that these AV interactions were not influenced by whether A and V were congruent or incongruent. In Experiment 3 we show that the AV interaction on N1 was absent when there was no anticipatory visual motion, indicating that the AV interaction only occurred when visual anticipatory motion preceded the sound. These results demonstrate that the visually induced speeding-up and suppression of auditory N1 amplitude reflect multisensory integrative mechanisms of AV events that crucially depend on whether vision predicts when the sound occurs.
Collapse
|
15
|
Davis C, Kislyuk D, Kim J, Sams M. The effect of viewing speech on auditory speech processing is different in the left and right hemispheres. Brain Res 2008; 1242:151-61. [PMID: 18538750 DOI: 10.1016/j.brainres.2008.04.077] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2007] [Revised: 04/30/2008] [Accepted: 04/30/2008] [Indexed: 10/22/2022]
Abstract
We used whole-head magnetoencephalograpy (MEG) to record changes in neuromagnetic N100m responses generated in the left and right auditory cortex as a function of the match between visual and auditory speech signals. Stimuli were auditory-only (AO) and auditory-visual (AV) presentations of /pi/, /ti/ and /vi/. Three types of intensity matched auditory stimuli were used: intact speech (Normal), frequency band filtered speech (Band) and speech-shaped white noise (Noise). The behavioural task was to detect the /vi/ syllables which comprised 12% of stimuli. N100m responses were measured to averaged /pi/ and /ti/ stimuli. Behavioural data showed that identification of the stimuli was faster and more accurate for Normal than for Band stimuli, and for Band than for Noise stimuli. Reaction times were faster for AV than AO stimuli. MEG data showed that in the left hemisphere, N100m to both AO and AV stimuli was largest for the Normal, smaller for Band and smallest for Noise stimuli. In the right hemisphere, Normal and Band AO stimuli elicited N100m responses of quite similar amplitudes, but N100m amplitude to Noise was about half of that. There was a reduction in N100m for the AV compared to the AO conditions. The size of this reduction for each stimulus type was same in the left hemisphere but graded in the right (being largest to the Normal, smaller to the Band and smallest to the Noise stimuli). The N100m decrease for the Normal stimuli was significantly larger in the right than in the left hemisphere. We suggest that the effect of processing visual speech seen in the right hemisphere likely reflects suppression of the auditory response based on AV cues for place of articulation.
Collapse
Affiliation(s)
- Chris Davis
- MARCS Auditory Laboratories, University of Western Sydney, Australia.
| | | | | | | |
Collapse
|
16
|
Mat Safri N, Murayama N, Hayashida Y, Igasaki T. Effects of concurrent visual tasks on cortico-muscular synchronization in humans. Brain Res 2007; 1155:81-92. [PMID: 17512919 DOI: 10.1016/j.brainres.2007.04.052] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2007] [Revised: 04/11/2007] [Accepted: 04/12/2007] [Indexed: 11/30/2022]
Abstract
To study the effects of external visual stimulation on motor cortex-muscle synchronization, coherence between electroencephalography (EEG) and electromyography (EMG) was measured in normal subjects under Before, Task (visual task: Ignore or Count, or arithmetic task) and After conditions. The control (Before and After) conditions required the subject to maintain first dorsal interosseous muscle contraction without visual stimulation. In the visual task, a random series of visual stimuli were displayed on a screen while the subjects maintained the muscle contraction. The subjects were asked to ignore the stimuli in the Ignore condition and to count certain stimuli in the Count condition. Also, in the arithmetic task, the subjects were asked to perform a simple subtraction. The EEG-EMG coherence found at C(3) site at 13-30 Hz (beta) was increased and sustained in magnitude during the Ignore and Count conditions, respectively. To examine the cause of the change of coherence, changes of EEG and EMG spectral power were computed for each frequency band. There was little change in the EMG spectral power in any frequency bands. While the spectral power of EEG unchanged in the beta band, it significantly increased and decreased in the range of 8-12 Hz and of 31-50 Hz, respectively, for both Ignore and Count conditions, not only at the C(3) site but at various sites as well. These results were in contrast to those obtained for the arithmetic task: the beta band EEG-EMG coherence was attenuated and the EEG spectral power at 4-7 Hz and at 31-50 Hz were significantly increased and decreased, respectively. As a conclusion, the present results are consistent with the idea that the enhanced 8-12 Hz/decreased 31-50 Hz oscillations affect strength of the beta band cortico-muscular synchronization by suppressing the visual processing.
Collapse
Affiliation(s)
- Norlaili Mat Safri
- Department of Graduate School of Science and Technology, Kumamoto University, Kurokami 2-39-1, Kumamoto, Japan
| | | | | | | |
Collapse
|
17
|
Miki K, Watanabe S, Honda Y, Nakamura M, Kakigi R. Effects of face contour and features on early occipitotemporal activity when viewing eye movement. Neuroimage 2007; 35:1624-35. [PMID: 17363279 DOI: 10.1016/j.neuroimage.2007.01.040] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2006] [Revised: 01/17/2007] [Accepted: 01/19/2007] [Indexed: 10/23/2022] Open
Abstract
We investigated whether early activity in the occipitotemporal region, corresponding to human MT/V5, is influenced by a face contour and/or features such as the mouth using magnetoencephalography (MEG). We used apparent motion as visual stimuli and compared four conditions as follows: (1) CDL: A schematic face consisting of a face Contour, two Dots and a horizontal Line (2) CD: The Contour and two Dots (3) DL: Two Dots and a horizontal Line and (4) D: Two Dots only. Subjects described a simple movement of dots for D, but eye movement for CDL, DL and CD, though movement modalities were the same through all conditions. We used a single equivalent current dipole (ECD) model within 145-220 ms after stimulus onset and estimated the location, dipole moment (strength) and peak latency. There were no significant differences in the peak latency of the estimated dipoles between each condition, but the activity was significantly stronger for CDL than for CD (p<0.05), DL (p<0.01), and D (p<0.01) in the right hemisphere, and DL and D (p<0.01) in the left. These results indicated that there is specific information processing for eye movements in the occipitotemporal region, the human MT/V5 homologue, and this activity was significantly influenced by whether movements appeared with the face contour and/or features, in other words, whether the eyes moved or not, even if the movement itself was the same.
Collapse
Affiliation(s)
- Kensaku Miki
- Department of Integrative Physiology, National Institute for Physiological Sciences, Okazaki, Aichi, Japan.
| | | | | | | | | |
Collapse
|
18
|
Hertrich I, Mathiak K, Lutzenberger W, Menning H, Ackermann H. Sequential audiovisual interactions during speech perception: A whole-head MEG study. Neuropsychologia 2007; 45:1342-54. [PMID: 17067640 DOI: 10.1016/j.neuropsychologia.2006.09.019] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2006] [Revised: 07/26/2006] [Accepted: 09/26/2006] [Indexed: 11/21/2022]
Abstract
Using whole-head magnetoencephalography (MEG), audiovisual (AV) interactions during speech perception (/ta/- and /pa/-syllables) were investigated in 20 subjects. Congruent AV events served as the 'standards' of an oddball design. The deviants encompassed incongruent /ta/-/pa/ configurations differing from the standards either in the acoustic or the visual domain. As an auditory non-speech control condition, the same video signals were synchronized with either one of two complex tones. As in natural speech, visual movement onset preceded acoustic signals by about 150 ms. First, the impact of visual information on auditorily evoked fields to non-speech sounds was determined. Larger facial movements (/pa/ versus /ta/) yielded enhanced early responses such as the M100 component, indicating, most presumably, anticipatory pre-activation of auditory cortex by visual motion cues. As a second step of analysis, mismatch fields (MMF) were calculated. Acoustic deviants elicited a typical MMF, peaking ca. 180 ms after stimulus onset, whereas visual deviants gave rise to later responses (220 ms) of a more posterior-medial source location. Finally, a late (275 ms), left-lateralized visually-induced MMF component, resembling the acoustic mismatch response, emerged during the speech condition, presumably reflecting phonetic/linguistic operations. There is mounting functional imaging evidence for an early impact of visual information on auditory cortical regions during speech perception. The present study suggests at least two successive AV interactions in association with syllable recognition tasks: early activation of auditory areas depending upon visual motion cues and a later speech-specific left-lateralized response mediated, conceivably, by backward-projections from multisensory areas.
Collapse
Affiliation(s)
- Ingo Hertrich
- Department of General Neurology, Hertie Institute for Clinical Brain Research, University of Tübingen, Germany.
| | | | | | | | | |
Collapse
|
19
|
Watanabe S, Miki K, Kakigi R. Mechanisms of face perception in humans: A magneto- and electro-encephalographic study. Neuropathology 2005; 25:8-20. [PMID: 15822814 DOI: 10.1111/j.1440-1789.2004.00603.x] [Citation(s) in RCA: 25] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
We have been studying the underlying mechanisms of face perception in humans using magneto- (MEG) and electro-encephalography (EEG) including (1) perception by viewing the static face, (2) differences in perception by viewing the eyes and whole face, (3) the face inversion effect, (4) the effect of gaze direction, (5) perception of eye motion, (6) perception of mouth motion, and (7) the interaction between auditory and visual stimuli related to the vowel sounds. In this review article, we mainly summarize our results obtained on 3, 5, and 6 above. With the presentation of both upright and inverted unfamiliar faces, the inferior temporal cortex (IT) centered on the fusiform gyrus, and the lateral temporal cortex (LT) near the superior temporal sulcus were activated simultaneously, but independently, between 140 and 200 ms post-stimulus. The right hemisphere IT and LT were both active in all subjects, and those in the left hemisphere in half of the subjects. Latencies with inverted faces relative to those with upright faces were longer in the right hemisphere, and shorter in the left hemisphere. Since the activated regions under upright and those under inverted face stimuli did not show a significant difference, we consider that differences in processing upright versus inverted faces are attributable to temporal processing differences rather than to processing of information by different brain regions. When viewing the motion of the mouth and eyes, a large clear MEG component, 1M (mean peak latency of approximately 160 ms), was elicited to both mouth and eye movement, and was generated mainly in the occipito-temporal border, at human MT/V5. The 1M to mouth movement and the 1M to eye movement showed no significant difference in amplitude or generator location. Therefore, our results indicate that human MT/V5 is active in the perception of both mouth and eye motion, and that the perception of movement of facial parts is probably processed similarly.
Collapse
Affiliation(s)
- Shoko Watanabe
- Department of Integrative Physiology, National Institute for Physiological Sciences, Okazaki, Japan
| | | | | |
Collapse
|
20
|
Kaneoke Y, Watanabe S, Kakigi R. Human Visual Processing as Revealed by Magnetoencephalography. INTERNATIONAL REVIEW OF NEUROBIOLOGY 2005; 68:197-222. [PMID: 16443015 DOI: 10.1016/s0074-7742(05)68008-7] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/11/2023]
Affiliation(s)
- Yoshiki Kaneoke
- Department of Integrative Physiology, National Institute for Physiological Sciences Myodaiji-cho, Okazaki, 444-8585, Japan
| | | | | |
Collapse
|
21
|
Besle J, Fort A, Delpuech C, Giard MH. Bimodal speech: early suppressive visual effects in human auditory cortex. Eur J Neurosci 2004; 20:2225-34. [PMID: 15450102 PMCID: PMC1885424 DOI: 10.1111/j.1460-9568.2004.03670.x] [Citation(s) in RCA: 232] [Impact Index Per Article: 11.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
Abstract
While everyone has experienced that seeing lip movements may improve speech perception, little is known about the neural mechanisms by which audiovisual speech information is combined. Event-related potentials (ERPs) were recorded while subjects performed an auditory recognition task among four different natural syllables randomly presented in the auditory (A), visual (V) or congruent bimodal (AV) condition. We found that: (i) bimodal syllables were identified more rapidly than auditory alone stimuli; (ii) this behavioural facilitation was associated with cross-modal [AV-(A+V)] ERP effects around 120-190 ms latency, expressed mainly as a decrease of unimodal N1 generator activities in the auditory cortex. This finding provides evidence for suppressive, speech-specific audiovisual integration mechanisms, which are likely to be related to the dominance of the auditory modality for speech perception. Furthermore, the latency of the effect indicates that integration operates at pre-representational stages of stimulus analysis, probably via feedback projections from visual and/or polymodal areas.
Collapse
|