1
|
McEwan J, Kritikos A, Zeljko M. Involvement of the superior colliculi in crossmodal correspondences. Atten Percept Psychophys 2024; 86:931-941. [PMID: 38418807 PMCID: PMC11062976 DOI: 10.3758/s13414-024-02866-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/11/2024] [Indexed: 03/02/2024]
Abstract
There is an increasing body of evidence suggesting that there are low-level perceptual processes involved in crossmodal correspondences. In this study, we investigate the involvement of the superior colliculi in three basic crossmodal correspondences: elevation/pitch, lightness/pitch, and size/pitch. Using a psychophysical design, we modulate visual input to the superior colliculus to test whether the superior colliculus is required for behavioural crossmodal congruency effects to manifest in an unspeeded multisensory discrimination task. In the elevation/pitch task, superior colliculus involvement is required for a behavioural elevation/pitch congruency effect to manifest in the task. In the lightness/pitch and size/pitch task, we observed a behavioural elevation/pitch congruency effect regardless of superior colliculus involvement. These results suggest that the elevation/pitch correspondence may be processed differently to other low-level crossmodal correspondences. The implications of a distributed model of crossmodal correspondence processing in the brain are discussed.
Collapse
Affiliation(s)
- John McEwan
- School of Psychology, The University of Queensland, St. Lucia, Queensland, 4072, Australia.
| | - Ada Kritikos
- School of Psychology, The University of Queensland, St. Lucia, Queensland, 4072, Australia
| | - Mick Zeljko
- School of Psychology, The University of Queensland, St. Lucia, Queensland, 4072, Australia
| |
Collapse
|
2
|
Barbosa Escobar F, Wang QJ. Inducing Novel Sound-Taste Correspondences via an Associative Learning Task. Cogn Sci 2024; 48:e13421. [PMID: 38500336 DOI: 10.1111/cogs.13421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2023] [Revised: 01/08/2024] [Accepted: 02/19/2024] [Indexed: 03/20/2024]
Abstract
The interest in crossmodal correspondences, including those involving sounds and involving tastes, has experienced rapid growth in recent years. However, the mechanisms underlying these correspondences are not well understood. In the present study (N = 302), we used an associative learning paradigm, based on previous literature using simple sounds with no consensual taste associations (i.e., square and triangle wave sounds at 200 Hz) and taste words (i.e., sweet and bitter), to test the influence of two potential mechanisms in establishing sound-taste correspondences and investigate whether either learning mechanism could give rise to new and long-lasting associations. Specifically, we examined an emotional mediation account (i.e., using sad and happy emoji facial expressions) and a transitive path (i.e., sound-taste correspondence being mediated by color, using red and black colored squares). The results revealed that the associative learning paradigm mapping the triangle wave tone with a happy emoji facial expression induced a novel crossmodal correspondence between this sound and the word sweet. Importantly, we found that this novel association was still present two months after the experimental learning paradigm. None of the other mappings, emotional or transitive, gave rise to any significant associations between sound and taste. These findings provide evidence that new crossmodal correspondences between sounds and tastes can be created by leveraging the affective connection between both dimensions, helping elucidate the mechanisms underlying these associations. Moreover, these findings reveal that these associations can last for several weeks after the experimental session through which they were induced.
Collapse
Affiliation(s)
- Francisco Barbosa Escobar
- Department of Food Science, Faculty of Science, University of Copenhagen
- Department of Marketing, Copenhagen Business School
| | - Qian Janice Wang
- Department of Food Science, Faculty of Science, University of Copenhagen
| |
Collapse
|
3
|
Dorsi J, Lacey S, Sathian K. Multisensory and lexical information in speech perception. Front Hum Neurosci 2024; 17:1331129. [PMID: 38259332 PMCID: PMC10800662 DOI: 10.3389/fnhum.2023.1331129] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Accepted: 12/11/2023] [Indexed: 01/24/2024] Open
Abstract
Both multisensory and lexical information are known to influence the perception of speech. However, an open question remains: is either source more fundamental to perceiving speech? In this perspective, we review the literature and argue that multisensory information plays a more fundamental role in speech perception than lexical information. Three sets of findings support this conclusion: first, reaction times and electroencephalographic signal latencies indicate that the effects of multisensory information on speech processing seem to occur earlier than the effects of lexical information. Second, non-auditory sensory input influences the perception of features that differentiate phonetic categories; thus, multisensory information determines what lexical information is ultimately processed. Finally, there is evidence that multisensory information helps form some lexical information as part of a phenomenon known as sound symbolism. These findings support a framework of speech perception that, while acknowledging the influential roles of both multisensory and lexical information, holds that multisensory information is more fundamental to the process.
Collapse
Affiliation(s)
- Josh Dorsi
- Department of Neurology, Penn State College of Medicine, Hershey, PA, United States
| | - Simon Lacey
- Department of Neurology, Penn State College of Medicine, Hershey, PA, United States
- Department of Neural and Behavioral Sciences, Penn State College of Medicine, Hershey, PA, United States
- Department of Psychology, Penn State Colleges of Medicine and Liberal Arts, Hershey, PA, United States
| | - K. Sathian
- Department of Neurology, Penn State College of Medicine, Hershey, PA, United States
- Department of Neural and Behavioral Sciences, Penn State College of Medicine, Hershey, PA, United States
- Department of Psychology, Penn State Colleges of Medicine and Liberal Arts, Hershey, PA, United States
| |
Collapse
|
4
|
Nalbantoğlu H, Hazır BM, Dövencioğlu DN. Selectively manipulating softness perception of materials through sound symbolism. Front Psychol 2024; 14:1323873. [PMID: 38259577 PMCID: PMC10801190 DOI: 10.3389/fpsyg.2023.1323873] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Accepted: 11/30/2023] [Indexed: 01/24/2024] Open
Abstract
Cross-modal interactions between auditory and haptic perception manifest themselves in language, such as sound symbolic words: crunch, splash, and creak. Several studies have shown strong associations between sound symbolic words, shapes (e.g., Bouba/Kiki effect), and materials. Here, we identified these material associations in Turkish sound symbolic words and then tested for their effect on softness perception. First, we used a rating task in a semantic differentiation method to extract the perceived softness dimensions from words and materials. We then tested whether Turkish onomatopoeic words can be used to manipulate the perceived softness of everyday materials such as honey, silk, or sand across different dimensions of softness. In the first preliminary study, we used 40 material videos and 29 adjectives in a rating task with a semantic differentiation method to extract the main softness dimensions. A principal component analysis revealed seven softness components, including Deformability, Viscosity, Surface Softness, and Granularity, in line with the literature. The second preliminary study used 27 onomatopoeic words and 21 adjectives in the same rating task. Again, the findings aligned with the literature, revealing dimensions such as Viscosity, Granularity, and Surface Softness. However, no factors related to Deformability were found due to the absence of sound symbolic words in this category. Next, we paired the onomatopoeic words and material videos based on their associations with each softness dimension. We conducted a new rating task, synchronously presenting material videos and spoken onomatopoeic words. We hypothesized that congruent word-video pairs would produce significantly higher ratings for dimension-related adjectives, while incongruent word-video pairs would decrease these ratings, and the ratings of unrelated adjectives would remain the same. Our results revealed that onomatopoeic words selectively alter the perceived material qualities, providing evidence and insight into the cross-modality of perceived softness.
Collapse
|
5
|
Chen L. Synesthetic Correspondence: An Overview. ADVANCES IN EXPERIMENTAL MEDICINE AND BIOLOGY 2024; 1437:101-119. [PMID: 38270856 DOI: 10.1007/978-981-99-7611-9_7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/26/2024]
Abstract
Intramodal and cross-modal perceptual grouping based on the spatial proximity and temporal closeness between multiple sensory stimuli, as an operational principle has built a coherent and meaningful representation of the multisensory event/object. To implement and investigate the cross-modal perceptual grouping, researchers have employed excellent paradigms of spatial/temporal ventriloquism and cross-modal dynamic capture and have revealed the conditional constraints as well as the functional facilitations among various correspondence of sensory properties, with featured behavioral evidence, computational framework as well as brain oscillation patterns. Typically, synesthetic correspondence as a special type of cross-modal correspondence can shape the efficiency and effect-size of cross-modal interaction. For example, factors such as pitch/loudness in the auditory dimension with size/brightness in the visual dimension could modulate the strength of the cross-modal temporal capture. The empirical behavioral findings, as well as psychophysical and neurophysiological evidence to address the cross-modal perceptual grouping and synesthetic correspondence, were summarized in this review. Finally, the potential applications (such as artificial synesthesia device) and how synesthetic correspondence interface with semantics (sensory linguistics), as well as the promising research questions in this field have been discussed.
Collapse
Affiliation(s)
- Lihan Chen
- School of Psychological and Cognitive Sciences, Peking University, Beijing, China.
- Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China.
- Key Laboratory of Machine Perception (Ministry of Education), Peking University, Beijing, China.
- National Key Laboratory of General Artificial Intelligence, Peking University, Beijing, China.
- National Engineering Laboratory for Big Data Analysis and Applications, Peking University, Beijing, China.
| |
Collapse
|
6
|
Barany DA, Lacey S, Matthews KL, Nygaard LC, Sathian K. Neural basis of sound-symbolic pseudoword-shape correspondences. Neuropsychologia 2023; 188:108657. [PMID: 37543139 PMCID: PMC10529692 DOI: 10.1016/j.neuropsychologia.2023.108657] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2023] [Revised: 06/23/2023] [Accepted: 08/02/2023] [Indexed: 08/07/2023]
Abstract
Non-arbitrary mapping between the sound of a word and its meaning, termed sound symbolism, is commonly studied through crossmodal correspondences between sounds and visual shapes, e.g., auditory pseudowords, like 'mohloh' and 'kehteh', are matched to rounded and pointed visual shapes, respectively. Here, we used functional magnetic resonance imaging (fMRI) during a crossmodal matching task to investigate the hypotheses that sound symbolism (1) involves language processing; (2) depends on multisensory integration; (3) reflects embodiment of speech in hand movements. These hypotheses lead to corresponding neuroanatomical predictions of crossmodal congruency effects in (1) the language network; (2) areas mediating multisensory processing, including visual and auditory cortex; (3) regions responsible for sensorimotor control of the hand and mouth. Right-handed participants (n = 22) encountered audiovisual stimuli comprising a simultaneously presented visual shape (rounded or pointed) and an auditory pseudoword ('mohloh' or 'kehteh') and indicated via a right-hand keypress whether the stimuli matched or not. Reaction times were faster for congruent than incongruent stimuli. Univariate analysis showed that activity was greater for the congruent compared to the incongruent condition in the left primary and association auditory cortex, and left anterior fusiform/parahippocampal gyri. Multivoxel pattern analysis revealed higher classification accuracy for the audiovisual stimuli when congruent than when incongruent, in the pars opercularis of the left inferior frontal (Broca's area), the left supramarginal, and the right mid-occipital gyri. These findings, considered in relation to the neuroanatomical predictions, support the first two hypotheses and suggest that sound symbolism involves both language processing and multisensory integration.
Collapse
Affiliation(s)
- Deborah A Barany
- Department of Kinesiology, University of Georgia and Augusta University/University of Georgia Medical Partnership, Athens, GA, 30602, USA
| | - Simon Lacey
- Department of Neurology, Penn State College of Medicine, Hershey, PA, 17033-0859, USA; Department of Neural & Behavioral Sciences, Penn State College of Medicine, Hershey, PA, 17033-0859, USA; Department of Psychology, Penn State College of Liberal Arts, University Park, PA, 16802, USA
| | - Kaitlyn L Matthews
- Department of Psychology, Emory University, Atlanta, GA, 30322, USA; Present address: Department of Psychological & Brain Sciences, Washington University in St. Louis, St. Louis, MO, 63130, USA
| | - Lynne C Nygaard
- Department of Psychology, Emory University, Atlanta, GA, 30322, USA
| | - K Sathian
- Department of Neurology, Penn State College of Medicine, Hershey, PA, 17033-0859, USA; Department of Neural & Behavioral Sciences, Penn State College of Medicine, Hershey, PA, 17033-0859, USA; Department of Psychology, Penn State College of Liberal Arts, University Park, PA, 16802, USA.
| |
Collapse
|
7
|
Barany DA, Lacey S, Matthews KL, Nygaard LC, Sathian K. Neural Basis Of Sound-Symbolic Pseudoword-Shape Correspondences. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.04.14.536865. [PMID: 37425853 PMCID: PMC10327042 DOI: 10.1101/2023.04.14.536865] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/11/2023]
Abstract
Non-arbitrary mapping between the sound of a word and its meaning, termed sound symbolism, is commonly studied through crossmodal correspondences between sounds and visual shapes, e.g., auditory pseudowords, like 'mohloh' and 'kehteh', are matched to rounded and pointed visual shapes, respectively. Here, we used functional magnetic resonance imaging (fMRI) during a crossmodal matching task to investigate the hypotheses that sound symbolism (1) involves language processing; (2) depends on multisensory integration; (3) reflects embodiment of speech in hand movements. These hypotheses lead to corresponding neuroanatomical predictions of crossmodal congruency effects in (1) the language network; (2) areas mediating multisensory processing, including visual and auditory cortex; (3) regions responsible for sensorimotor control of the hand and mouth. Right-handed participants ( n = 22) encountered audiovisual stimuli comprising a simultaneously presented visual shape (rounded or pointed) and an auditory pseudoword ('mohloh' or 'kehteh') and indicated via a right-hand keypress whether the stimuli matched or not. Reaction times were faster for congruent than incongruent stimuli. Univariate analysis showed that activity was greater for the congruent compared to the incongruent condition in the left primary and association auditory cortex, and left anterior fusiform/parahippocampal gyri. Multivoxel pattern analysis revealed higher classification accuracy for the audiovisual stimuli when congruent than when incongruent, in the pars opercularis of the left inferior frontal (Broca's area), the left supramarginal, and the right mid-occipital gyri. These findings, considered in relation to the neuroanatomical predictions, support the first two hypotheses and suggest that sound symbolism involves both language processing and multisensory integration. HIGHLIGHTS fMRI investigation of sound-symbolic correspondences between auditory pseudowords and visual shapesFaster reaction times for congruent than incongruent audiovisual stimuliGreater activation in auditory and visual cortices for congruent stimuliHigher classification accuracy for congruent stimuli in language and visual areasSound symbolism involves language processing and multisensory integration.
Collapse
Affiliation(s)
- Deborah A. Barany
- Department of Kinesiology, University of Georgia and Augusta University/University of Georgia Medical Partnership, Athens, GA, 30602, USA
| | - Simon Lacey
- Department of Neurology, Penn State Colleges of Medicine and Liberal Arts, Hershey, PA 17033-0859, USA
- Department of Neural & Behavioral Sciences, Penn State Colleges of Medicine and Liberal Arts, Hershey, PA 17033-0859, USA
- Department of Psychology, Penn State Colleges of Medicine and Liberal Arts, Hershey, PA 17033-0859, USA
| | - Kaitlyn L. Matthews
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
- Present address: Department of Psychological & Brain Sciences, Washington University in St. Louis, St. Louis, MO 63130
| | - Lynne C. Nygaard
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - K. Sathian
- Department of Neurology, Penn State Colleges of Medicine and Liberal Arts, Hershey, PA 17033-0859, USA
- Department of Neural & Behavioral Sciences, Penn State Colleges of Medicine and Liberal Arts, Hershey, PA 17033-0859, USA
- Department of Psychology, Penn State Colleges of Medicine and Liberal Arts, Hershey, PA 17033-0859, USA
| |
Collapse
|
8
|
Chen YC, Huang PC. Examining the automaticity and symmetry of sound-shape correspondences. Front Psychol 2023; 14:1172946. [PMID: 37342641 PMCID: PMC10277733 DOI: 10.3389/fpsyg.2023.1172946] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2023] [Accepted: 05/16/2023] [Indexed: 06/23/2023] Open
Abstract
Introduction A classic example of sound-shape correspondences is the mapping of the vowel /i/ with angular patterns and the vowel /u/ with rounded patterns. Such crossmodal correspondences have been reliably reported when tested in explicit matching tasks. Nevertheless, it remains unclear whether such sound-shape correspondences automatically occur and bidirectionally modulate people's perception. We address this question by adopting the explicit matching task and two implicit tasks. Methods In Experiment 1, we examined the sound-shape correspondences using the implicit association test (IAT), in which the sounds and shapes were both task-relevant, followed by an explicit matching task. In Experiments 2 and 3, we adopted the speeded classification task; when the target was a sound (or shape), a task-irrelevant shape (or sound) that was congruent or incongruent to the target was simultaneously presented. In addition, the participants performed the explicit matching task either before or after the speeded classification task. Results and Discussion The congruency effect was more pronounced in the IAT than in the speeded classification task; in addition, a bin analysis of RTs revealed that the congruency effect took time to develop. These findings suggest that the sound-shape correspondences were not completely automatic. The magnitude and onset of visual and auditory congruency effects were comparable, suggesting that the crossmodal modulations were symmetrical. Taken together, the sound-shape correspondences appeared not to be completely automatic, but their modulation was bidirectionally symmetrical once it occurred.
Collapse
Affiliation(s)
- Yi-Chuan Chen
- Department of Medicine, MacKay Medical College, New Taipei City, Taiwan
| | - Pi-Chun Huang
- Department of Psychology, National Cheng Kung University, Tainan, Taiwan
| |
Collapse
|
9
|
Tkacheva L, Flaksman M, Sedelkina Y, Lavitskaya Y, Nasledov A, Korotaevskaya E. Neural Indicators of Visual Andauditory Recognition of Imitative Words on Different De-Iconization Stages. Brain Sci 2023; 13:brainsci13040681. [PMID: 37190646 DOI: 10.3390/brainsci13040681] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2023] [Revised: 04/16/2023] [Accepted: 04/17/2023] [Indexed: 05/17/2023] Open
Abstract
The research aims to reveal neural indicators of recognition for iconic words and the possible cross-modal multisensory integration behind this process. The goals of this research are twofold: (1) to register event-related potentials (ERP) in the brain in the process of visual and auditory recognition of Russian imitative words on different de-iconization stages; and (2) to establish whether differences in the brain activity arise while processing visual and auditory stimuli of different nature. Sound imitative (onomatopoeic, mimetic, and ideophonic) words are words with iconic correlation between form and meaning (iconicity being a relationship of resemblance). Russian adult participants (n = 110) were presented with 15 stimuli both visually and auditorily. The stimuli material was equally distributed into three groups according to the criterion of (historical) iconicity loss: five explicit sound imitative (SI) words, five implicit SI words and five non-SI words. It was established that there was no statistically significant difference between visually presented explicit or implicit SI words and non-SI words respectively. However, statistically significant differences were registered for auditorily presented explicit SI words in contrast to implicit SI words in the N400 ERP component, as well as implicit SI words in contrast to non-SI words in the P300 ERP component. We thoroughly analyzed the integrative brain activity in response to explicit IS words and compared it to that in response to implicit SI and non-SI words presented auditorily. The data yielded by this analysis showed the N400 ERP component was more prominent during the recognition process of the explicit SI words received from the central channels (specifically Cz). We assume that these results indicate a specific brain response associated with directed attention in the process of performing cognitive decision making tasks regarding explicit and implicit SI words presented auditorily. This may reflect a higher level of cognitive complexity in identifying this type of stimuli considering the experimental task challenges that may involve cross-modal integration process.
Collapse
Affiliation(s)
- Liubov Tkacheva
- Department of Pedagogy and Pedagogical Psychology, Saint Petersburg State University, 199034 Saint Petersburg, Russia
| | - Maria Flaksman
- Department for English and American Studies, Ludwig Maximilian University, 80799 München, Germany
| | - Yulia Sedelkina
- Department of Foreign Languages and Linguo-Didactics, Saint Petersburg State University, 199034 Saint Petersburg, Russia
| | - Yulia Lavitskaya
- Department of Foreign Languages and Linguo-Didactics, Saint Petersburg State University, 199034 Saint Petersburg, Russia
| | - Andrey Nasledov
- Department of Pedagogy and Pedagogical Psychology, Saint Petersburg State University, 199034 Saint Petersburg, Russia
| | - Elizaveta Korotaevskaya
- Department of Clinical Psychology, Saint Petersburg State University, 199034 Saint Petersburg, Russia
| |
Collapse
|
10
|
Sciortino P, Kayser C. Steady state visual evoked potentials reveal a signature of the pitch-size crossmodal association in visual cortex. Neuroimage 2023; 273:120093. [PMID: 37028733 DOI: 10.1016/j.neuroimage.2023.120093] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2023] [Revised: 03/31/2023] [Accepted: 04/04/2023] [Indexed: 04/08/2023] Open
Abstract
Crossmodal correspondences describe our tendency to associate sensory features from different modalities with each other, such as the pitch of a sound with the size of a visual object. While such crossmodal correspondences (or associations) are described in many behavioural studies their neurophysiological correlates remain unclear. Under the current working model of multisensory perception both a low- and a high-level account seem plausible. That is, the neurophysiological processes shaping these associations could commence in low-level sensory regions, or may predominantly emerge in high-level association regions of semantic and object identification networks. We exploited steady-state visual evoked potentials (SSVEP) to directly probe this question, focusing on the associations between pitch and the visual features of size, hue or chromatic saturation. We found that SSVEPs over occipital regions are sensitive to the congruency between pitch and size, and a source analysis pointed to an origin around primary visual cortices. We speculate that this signature of the pitch-size association in low-level visual cortices reflects the successful pairing of congruent visual and acoustic object properties and may contribute to establishing causal relations between multisensory objects. Besides this, our study also provides a paradigm can be exploited to study other crossmodal associations involving visual stimuli in the future.
Collapse
|
11
|
Resolving the bouba-kiki effect enigma by rooting iconic sound symbolism in physical properties of round and spiky objects. Sci Rep 2022; 12:19172. [PMID: 36357511 PMCID: PMC9649795 DOI: 10.1038/s41598-022-23623-w] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2022] [Accepted: 11/02/2022] [Indexed: 11/11/2022] Open
Abstract
The "bouba-kiki effect", where "bouba" is perceived round and "kiki" spiky, remains a puzzling enigma. We solve it by combining mathematical findings largely unknown in the field, with computational models and novel experimental evidence. We reveal that this effect relies on two acoustic cues: spectral balance and temporal continuity. We demonstrate that it is not speech-specific but rather rooted in physical properties of objects, creating audiovisual regularities in the environment. Round items are mathematically bound to produce, when hitting or rolling on a surface, lower-frequency spectra and more continuous sounds than same-size spiky objects. Finally, we show that adults are sensitive to such regularities. Hence, intuitive physics impacts language perception and possibly language acquisition and evolution too.
Collapse
|
12
|
Shen YC, Chen YC, Huang PC. Seeing Sounds: The Role of Vowels and Consonants in Crossmodal Correspondences. Iperception 2022; 13:20416695221084724. [PMID: 35321530 PMCID: PMC8935407 DOI: 10.1177/20416695221084724] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2021] [Accepted: 02/11/2022] [Indexed: 11/17/2022] Open
Abstract
Crossmodal correspondences refer to the fact that certain domains of features in different sensory modalities are associated with each other. Here, we investigated the crossmodal correspondences between speech sounds and visual shapes. Specifically, we tested whether the classification dimensions of English vowels (front–central–back) and consonants (voiced–voiceless, sonorant–obstruent, and stop–continuant) correspond to visual shapes along a bipolar rounded–angular dimension. We adapted eighteen meaningless pseudowords from a previous study that corresponded to either the round or the sharp concept. On each trial, the participants heard one of the pseudowords and saw a rounded shape and an angular shape presented side-by-side on the monitor. Participants judged which shape provided a better match to the spoken pseudoword. A logistic regression was conducted in order to elucidate the effectiveness of classification dimensions of phonemes when predicting variations in the sound–shape matchings. The results demonstrated that the sound–shape matchings were predictable using front–central–back dimensions of vowels, and voiced–voiceless and stop–continuant dimensions of consonants. Hence, we verified that sound–shape matching is underpinned by contrasting dimensions in both vowels and consonants, therefore demonstrating crossmodal correspondences at the phonetic level.
Collapse
Affiliation(s)
- Yang-Chen Shen
- Department of Psychology, National Cheng Kung University, Tainan
| | - Yi-Chuan Chen
- Department of Medicine, Mackay Medical College, New Taipei City
| | - Pi-Chun Huang
- Department of Psychology, National Cheng Kung University, Tainan
| |
Collapse
|
13
|
Gold R, Klein D, Segal O. The Bouba-Kiki Effect in Children With Childhood Apraxia of Speech. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:43-52. [PMID: 34890216 DOI: 10.1044/2021_jslhr-21-00070] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
PURPOSE The bouba-kiki (BK) effect refers to associations between visual shapes and auditory pseudonames. Thus, when tested, people tend to associate the pseudowords bouba and kiki with round or spiky shapes, respectively. This association requires cross-modal sensory integration. The ability to integrate information from different sensory modalities is crucial for speech development. A clinical population that may be impaired in cross-modal sensory integration is children with childhood apraxia of speech (CAS). The purpose of this study was to examine the involvement of cross-modal sensory integration in children with (CAS). METHOD The BK effect was assessed in participants with CAS (n = 18) and two control groups: One control group was composed of children with developmental language disorder (DLD), also termed specific language impairment (n = 15), and a second group included typically developing (TD) children (n = 22). The children were presented with 14 pairs of novel visual displays and nonwords. All the children were asked to state which shape and nonword correspond to one another. In addition, background cognitive (Leiter-3) and language measures (Hebrew PLS-4) were determined for all children. RESULTS Children in the CAS group were less successful in associating between visual shapes and corresponding auditory pseudonames (e.g., associating the spoken word "bouba" with a round shape; the spoken word "kiki" with a spiky shape). Thus, children with CAS demonstrated a statistically significant reduced BK effect compared with participants with TD and participants with DLD. No significant difference was found between the TD group and the DLD group. CONCLUSIONS The reduced BK effect in children with CAS supports the notion that cross-modal sensory integration may be altered in these children. Cross-modal sensory integration is the basis for speech production. Thus, difficulties in sensory integration may contribute to speech difficulties in CAS.
Collapse
Affiliation(s)
- Rinat Gold
- Department of Communication Disorders, Sackler Faculty of Medicine, Tel Aviv University, Israel
| | - Dina Klein
- Department of Communication Disorders, Sackler Faculty of Medicine, Tel Aviv University, Israel
| | - Osnat Segal
- Department of Communication Disorders, Sackler Faculty of Medicine, Tel Aviv University, Israel
| |
Collapse
|
14
|
McCormick K, Lacey S, Stilla R, Nygaard LC, Sathian K. Neural Basis of the Sound-Symbolic Crossmodal Correspondence Between Auditory Pseudowords and Visual Shapes. Multisens Res 2021; 35:29-78. [PMID: 34384048 PMCID: PMC9196751 DOI: 10.1163/22134808-bja10060] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2020] [Accepted: 07/17/2021] [Indexed: 11/19/2022]
Abstract
Sound symbolism refers to the association between the sounds of words and their meanings, often studied using the crossmodal correspondence between auditory pseudowords, e.g., 'takete' or 'maluma', and pointed or rounded visual shapes, respectively. In a functional magnetic resonance imaging study, participants were presented with pseudoword-shape pairs that were sound-symbolically congruent or incongruent. We found no significant congruency effects in the blood oxygenation level-dependent (BOLD) signal when participants were attending to visual shapes. During attention to auditory pseudowords, however, we observed greater BOLD activity for incongruent compared to congruent audiovisual pairs bilaterally in the intraparietal sulcus and supramarginal gyrus, and in the left middle frontal gyrus. We compared this activity to independent functional contrasts designed to test competing explanations of sound symbolism, but found no evidence for mediation via language, and only limited evidence for accounts based on multisensory integration and a general magnitude system. Instead, we suggest that the observed incongruency effects are likely to reflect phonological processing and/or multisensory attention. These findings advance our understanding of sound-to-meaning mapping in the brain.
Collapse
Affiliation(s)
- Kelly McCormick
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - Simon Lacey
- Department of Neurology, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
- Department of Neural and Behavioral Sciences, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
| | - Randall Stilla
- Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
| | - Lynne C. Nygaard
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - K. Sathian
- Department of Neurology, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
- Department of Neural and Behavioral Sciences, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
- Department of Psychology, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
| |
Collapse
|
15
|
Kitada R, Kwon J, Doizaki R, Nakagawa E, Tanigawa T, Kajimoto H, Sadato N, Sakamoto M. Brain networks underlying the processing of sound symbolism related to softness perception. Sci Rep 2021; 11:7399. [PMID: 33795716 PMCID: PMC8016892 DOI: 10.1038/s41598-021-86328-6] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2020] [Accepted: 03/15/2021] [Indexed: 01/02/2023] Open
Abstract
Unlike the assumption of modern linguistics, there is non-arbitrary association between sound and meaning in sound symbolic words. Neuroimaging studies have suggested the unique contribution of the superior temporal sulcus to the processing of sound symbolism. However, because these findings are limited to the mapping between sound symbolism and visually presented objects, the processing of sound symbolic information may also involve the sensory-modality dependent mechanisms. Here, we conducted a functional magnetic resonance imaging experiment to test whether the brain regions engaged in the tactile processing of object properties are also involved in mapping sound symbolic information with tactually perceived object properties. Thirty-two healthy subjects conducted a matching task in which they judged the congruency between softness perceived by touch and softness associated with sound symbolic words. Congruency effect was observed in the orbitofrontal cortex, inferior frontal gyrus, insula, medial superior frontal gyrus, cingulate gyrus, and cerebellum. This effect in the insula and medial superior frontal gyri was overlapped with softness-related activity that was separately measured in the same subjects in the tactile experiment. These results indicate that the insula and medial superior frontal gyrus play a role in processing sound symbolic information and relating it to the tactile softness information.
Collapse
Affiliation(s)
- Ryo Kitada
- Division of Psychology, School of Social Sciences, Nanyang Technological University, 48 Nanyang Avenue, Singapore, 639818, Singapore.
- Faculty of Intercultural Studies, Kobe University, 1-2-1 TsuruKabuto, Nada-ku, Kobe, 657-8501, Japan.
| | - Jinhwan Kwon
- Kyoto University of Education, Fukakusa-Fujimori-cho 1, Fushimi-ku, Kyoto, 612-8522, Japan
| | - Ryuichi Doizaki
- Department of Informatics, Graduate School of Informatics and Engineering, The University of Electro-Communications, 1-5-1 Chofugaoka, Chofu, Tokyo, 182-8585, Japan
| | - Eri Nakagawa
- National Institute for Physiological Sciences, Nishigonaka 38, Myodaiji-cho, Okazaki, 444-8585, Japan
| | - Tsubasa Tanigawa
- National Institute for Physiological Sciences, Nishigonaka 38, Myodaiji-cho, Okazaki, 444-8585, Japan
- The Graduate University for Advanced Studies (SOKENDAI), Shonan Village, Hayama, Kanagawa, 240-0193, Japan
| | - Hiroyuki Kajimoto
- Department of Informatics, Graduate School of Informatics and Engineering, The University of Electro-Communications, 1-5-1 Chofugaoka, Chofu, Tokyo, 182-8585, Japan
| | - Norihiro Sadato
- National Institute for Physiological Sciences, Nishigonaka 38, Myodaiji-cho, Okazaki, 444-8585, Japan
- The Graduate University for Advanced Studies (SOKENDAI), Shonan Village, Hayama, Kanagawa, 240-0193, Japan
| | - Maki Sakamoto
- Department of Informatics, Graduate School of Informatics and Engineering, The University of Electro-Communications, 1-5-1 Chofugaoka, Chofu, Tokyo, 182-8585, Japan
| |
Collapse
|
16
|
Kwak Y, Nam H, Kim HW, Kim CY. Cross-Modal Correspondence Between Speech Sound and Visual Shape Influencing Perceptual Representation of Shape: the Role of Articulation and Pitch. Multisens Res 2020; 33:569-598. [PMID: 32083558 DOI: 10.1163/22134808-20191330] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2018] [Accepted: 10/21/2019] [Indexed: 11/19/2022]
Abstract
Cross-modal correspondence is the tendency to systematically map stimulus features across sensory modalities. The current study explored cross-modal correspondence between speech sound and shape (Experiment 1), and whether such association can influence shape representation (Experiment 2). For the purpose of closely examining the role of the two factors - articulation and pitch - combined in speech acoustics, we generated two sets of 25 vowel stimuli - pitch-varying and pitch-constant sets. Both sets were generated by manipulating articulation - frontness and height of the tongue body's positions - but differed in terms of whether pitch varied among the sounds within the same set. In Experiment 1, participants made a forced choice between a round and a spiky shape to indicate the shape better associated with each sound. Results showed that shape choice was modulated according to both articulation and pitch, and we therefore concluded that both factors play significant roles in sound-shape correspondence. In Experiment 2, participants reported their subjective experience of shape accompanied by vowel sounds by adjusting an ambiguous shape in the response display. We found that sound-shape correspondence exerts an effect on shape representation by modulating audiovisual interaction, but only in the case of pitch-varying sounds. Therefore, pitch information within vowel acoustics plays the leading role in sound-shape correspondence influencing shape representation. Taken together, our results suggest the importance of teasing apart the roles of articulation and pitch for understanding sound-shape correspondence.
Collapse
Affiliation(s)
- Yuna Kwak
- 1Department of Psychology, Korea University, Seoul 02841, Korea
| | - Hosung Nam
- 2Department of English Language and Literature, Korea University, Seoul 02841, Korea.,3Haskins Laboratories, New Haven, CT 06511, USA
| | - Hyun-Woong Kim
- 1Department of Psychology, Korea University, Seoul 02841, Korea
| | - Chai-Youn Kim
- 1Department of Psychology, Korea University, Seoul 02841, Korea
| |
Collapse
|
17
|
Hus Y, Segal O. Functional Communication Profiles of Children and Youth with Autism: A Roadmap to Therapeutic and Educational Interventions. Folia Phoniatr Logop 2020; 73:205-221. [PMID: 32937618 DOI: 10.1159/000510056] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2020] [Accepted: 07/09/2020] [Indexed: 11/19/2022] Open
Abstract
BACKGROUND Autism spectrum disorder (ASD) entails varied developmental pathways along the entire lifespan, demanding early and ongoing diverse and responsive interventions to children's needs. This study examined in situ education and development attained by children and youth with ASD in a school with a therapeutic and educational curriculum. OBJECTIVES (1) Construct individual communication profiles in educational and developmental aspects. (2) Examine for associations between variables. (3) Demonstrate the clinical and educational utility of including cognitive-linguistic integrative variables. (4) Showcase how the profiles guide interventions tailored to students' individual needs. METHODS Functional communication profiles (FCPs) of 21 students, aged 5.0-16.8 years, mostly from bilingual middle-high socioeconomic status families, were constructed with input from their educational and therapeutic staff. Students' performance was examined with an array of instruments and tasks, including person and clock drawings, false belief, and bouba-kiki metaphor screens. Qualitative and quantitative analyses were performed to uncover associations, weaknesses, and strengths. RESULTS The profiles revealed associations between cognitive, linguistic, social, and educational abilities, indicating that some abilities from different domains tend to co-occur. CONCLUSIONS Including cognitive linguistic integrative variables was a novelty that revealed additional aspects of the children's abilities. Staff feedback confirmed the utility of FCPs in providing "a roadmap" to needed individual and common curriculum adjustments.
Collapse
|
18
|
Lacey S, Jamal Y, List SM, McCormick K, Sathian K, Nygaard LC. Stimulus Parameters Underlying Sound-Symbolic Mapping of Auditory Pseudowords to Visual Shapes. Cogn Sci 2020; 44:e12883. [PMID: 32909637 PMCID: PMC7896554 DOI: 10.1111/cogs.12883] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2018] [Revised: 06/06/2020] [Accepted: 07/01/2020] [Indexed: 12/12/2022]
Abstract
Sound symbolism refers to non-arbitrary mappings between the sounds of words and their meanings and is often studied by pairing auditory pseudowords such as "maluma" and "takete" with rounded and pointed visual shapes, respectively. However, it is unclear what auditory properties of pseudowords contribute to their perception as rounded or pointed. Here, we compared perceptual ratings of the roundedness/pointedness of large sets of pseudowords and shapes to their acoustic and visual properties using a novel application of representational similarity analysis (RSA). Representational dissimilarity matrices (RDMs) of the auditory and visual ratings of roundedness/pointedness were significantly correlated crossmodally. The auditory perceptual RDM correlated significantly with RDMs of spectral tilt, the temporal fast Fourier transform (FFT), and the speech envelope. Conventional correlational analyses showed that ratings of pseudowords transitioned from rounded to pointed as vocal roughness (as measured by the harmonics-to-noise ratio, pulse number, fraction of unvoiced frames, mean autocorrelation, shimmer, and jitter) increased. The visual perceptual RDM correlated significantly with RDMs of global indices of visual shape (the simple matching coefficient, image silhouette, image outlines, and Jaccard distance). Crossmodally, the RDMs of the auditory spectral parameters correlated weakly but significantly with those of the global indices of visual shape. Our work establishes the utility of RSA for analysis of large stimulus sets and offers novel insights into the stimulus parameters underlying sound symbolism, showing that sound-to-shape mapping is driven by acoustic properties of pseudowords and suggesting audiovisual cross-modal correspondence as a basis for language users' sensitivity to this type of sound symbolism.
Collapse
Affiliation(s)
- Simon Lacey
- Department of Neurology, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
- Department of Neural & Behavioral Sciences, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
- Department of Neurology, Emory University, Atlanta, GA 30322, USA
| | - Yaseen Jamal
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - Sara M. List
- Department of Neurology, Emory University, Atlanta, GA 30322, USA
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - Kelly McCormick
- Department of Neurology, Emory University, Atlanta, GA 30322, USA
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - K. Sathian
- Department of Neurology, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
- Department of Neural & Behavioral Sciences, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
- Department of Psychology, Milton S. Hershey Medical Center, Penn State College of Medicine, Hershey, PA 17033-0859, USA
- Department of Neurology, Emory University, Atlanta, GA 30322, USA
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| | - Lynne C. Nygaard
- Department of Psychology, Emory University, Atlanta, GA 30322, USA
| |
Collapse
|