1
|
Liao C, Sawayama M, Xiao B. Probing the link between vision and language in material perception using psychophysics and unsupervised learning. PLoS Comput Biol 2024; 20:e1012481. [PMID: 39361707 PMCID: PMC11478833 DOI: 10.1371/journal.pcbi.1012481] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2024] [Revised: 10/15/2024] [Accepted: 09/11/2024] [Indexed: 10/05/2024] Open
Abstract
We can visually discriminate and recognize a wide range of materials. Meanwhile, we use language to describe what we see and communicate relevant information about the materials. Here, we investigate the relationship between visual judgment and language expression to understand how visual features relate to semantic representations in human cognition. We use deep generative models to generate images of realistic materials. Interpolating between the generative models enables us to systematically create material appearances in both well-defined and ambiguous categories. Using these stimuli, we compared the representations of materials from two behavioral tasks: visual material similarity judgments and free-form verbal descriptions. Our findings reveal a moderate but significant correlation between vision and language on a categorical level. However, analyzing the representations with an unsupervised alignment method, we discover structural differences that arise at the image-to-image level, especially among ambiguous materials morphed between known categories. Moreover, visual judgments exhibit more individual differences compared to verbal descriptions. Our results show that while verbal descriptions capture material qualities on the coarse level, they may not fully convey the visual nuances of material appearances. Analyzing the image representation of materials obtained from various pre-trained deep neural networks, we find that similarity structures in human visual judgments align more closely with those of the vision-language models than purely vision-based models. Our work illustrates the need to consider the vision-language relationship in building a comprehensive model for material perception. Moreover, we propose a novel framework for evaluating the alignment and misalignment between representations from different modalities, leveraging information from human behaviors and computational models.
Collapse
Affiliation(s)
- Chenxi Liao
- American University, Department of Neuroscience, Washington DC, United States of America
| | - Masataka Sawayama
- The University of Tokyo, Graduate School of Information Science and Technology, Tokyo, Japan
| | - Bei Xiao
- American University, Department of Computer Science, Washington DC, United States of America
| |
Collapse
|
2
|
Kılıç F, Dövencioğlu D. Visual softness perception can be manipulated through exploratory procedures. Perception 2024; 53:674-687. [PMID: 39053476 DOI: 10.1177/03010066241261772] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/27/2024]
Abstract
Both visual and haptic softness perception have recently been shown to have multiple dimensions, such as deformability, granularity, fluidity, surface softness, and roughness. During haptic exploration, people adjust their hand motions (exploratory procedures, EPs) based on the material qualities of the object and the particular information they intend to acquire. Some of these EPs are also shown to be associated with perceived softness dimensions, for example, stroking a silk blouse or applying pressure to a pillow. Here, we aimed to investigate whether we can manipulate observers' judgments about softness attributes through exposure to videos of others performing various EPs on everyday soft materials. In two experiments, participants watched two videos of the same material: one with a corresponding EP and the other without correspondence; then, they judged these materials based on 12 softness-related adjectives (semantic differentiation method). The results of the second experiment suggested that when the EP is congruent with the dimension from which the material is chosen, the ratings for the adjectives from the same dimension are higher than the incongruent EP. This study provides evidence that participants can assess material properties from optic and mechanical cues without needing haptic signals. Additionally, our findings indicate that manipulating the hand motion can selectively facilitate material-related judgments.
Collapse
|
3
|
Liao C, Sawayama M, Xiao B. Probing the Link Between Vision and Language in Material Perception Using Psychophysics and Unsupervised Learning. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.01.25.577219. [PMID: 38328102 PMCID: PMC10849714 DOI: 10.1101/2024.01.25.577219] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/09/2024]
Abstract
We can visually discriminate and recognize a wide range of materials. Meanwhile, we use language to express our subjective understanding of visual input and communicate relevant information about the materials. Here, we investigate the relationship between visual judgment and language expression in material perception to understand how visual features relate to semantic representations. We use deep generative networks to construct an expandable image space to systematically create materials of well-defined and ambiguous categories. From such a space, we sampled diverse stimuli and compared the representations of materials from two behavioral tasks: visual material similarity judgments and free-form verbal descriptions. Our findings reveal a moderate but significant correlation between vision and language on a categorical level. However, analyzing the representations with an unsupervised alignment method, we discover structural differences that arise at the image-to-image level, especially among materials morphed between known categories. Moreover, visual judgments exhibit more individual differences compared to verbal descriptions. Our results show that while verbal descriptions capture material qualities on the coarse level, they may not fully convey the visual features that characterize the material's optical properties. Analyzing the image representation of materials obtained from various pre-trained data-rich deep neural networks, we find that human visual judgments' similarity structures align more closely with those of the text-guided visual-semantic model than purely vision-based models. Our findings suggest that while semantic representations facilitate material categorization, non-semantic visual features also play a significant role in discriminating materials at a finer level. This work illustrates the need to consider the vision-language relationship in building a comprehensive model for material perception. Moreover, we propose a novel framework for quantitatively evaluating the alignment and misalignment between representations from different modalities, leveraging information from human behaviors and computational models.
Collapse
Affiliation(s)
- Chenxi Liao
- American University, Department of Neuroscience, Washington DC, 20016, USA
| | - Masataka Sawayama
- The University of Tokyo, Graduate School of Information Science and Technology, Tokyo, 113-0033, Japan
| | - Bei Xiao
- American University, Department of Computer Science, Washington, DC, 20016, USA
| |
Collapse
|
4
|
Mete M, Jeong H, Wang WD, Paik J. SORI: A softness-rendering interface to unravel the nature of softness perception. Proc Natl Acad Sci U S A 2024; 121:e2314901121. [PMID: 38466880 PMCID: PMC10990105 DOI: 10.1073/pnas.2314901121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Accepted: 02/02/2024] [Indexed: 03/13/2024] Open
Abstract
Tactile perception of softness serves a critical role in the survival, well-being, and social interaction among various species, including humans. This perception informs activities from food selection in animals to medical palpation for disease detection in humans. Despite its fundamental importance, a comprehensive understanding of how softness is neurologically and cognitively processed remains elusive. Previous research has demonstrated that the somatosensory system leverages both cutaneous and kinesthetic cues for the sensation of softness. Factors such as contact area, depth, and force play a particularly critical role in sensations experienced at the fingertips. Yet, existing haptic technologies designed to explore this phenomenon are limited, as they often couple force and contact area, failing to provide a real-world experience of softness perception. Our research introduces the softness-rendering interface (SORI), a haptic softness display designed to bridge this knowledge gap. Unlike its predecessors, SORI has the unique ability to decouple contact area and force, thereby allowing for a quantitative representation of softness sensations at the fingertips. Furthermore, SORI incorporates individual physical fingertip properties and model-based softness cue estimation and mapping to provide a highly personalized experience. Utilizing this method, SORI quantitatively replicates the sensation of softness on stationary, dynamic, homogeneous, and heterogeneous surfaces. We demonstrate that SORI accurately renders the surfaces of both virtual and daily objects, thereby presenting opportunities across a range of fields, from teleoperation to medical technology. Finally, our proposed method and SORI will expedite psychological and neuroscience research to unlock the nature of softness perception.
Collapse
Affiliation(s)
- Mustafa Mete
- Reconfigurable Robotics Laboratory, Institute of Mechanical Engineering, School of Engineering, École Polytechnique Fédérale de Lausanne, LausanneCH 1005, Switzerland
| | - Haewon Jeong
- Soft Robotics Laboratory, Department of Mechanical Engineering, College of Engineering, Hanyang University, Seoul04763, Republic of Korea
| | - Wei Dawid Wang
- Soft Robotics Laboratory, Department of Mechanical Engineering, College of Engineering, Hanyang University, Seoul04763, Republic of Korea
| | - Jamie Paik
- Reconfigurable Robotics Laboratory, Institute of Mechanical Engineering, School of Engineering, École Polytechnique Fédérale de Lausanne, LausanneCH 1005, Switzerland
| |
Collapse
|
5
|
Nalbantoğlu H, Hazır BM, Dövencioğlu DN. Selectively manipulating softness perception of materials through sound symbolism. Front Psychol 2024; 14:1323873. [PMID: 38259577 PMCID: PMC10801190 DOI: 10.3389/fpsyg.2023.1323873] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Accepted: 11/30/2023] [Indexed: 01/24/2024] Open
Abstract
Cross-modal interactions between auditory and haptic perception manifest themselves in language, such as sound symbolic words: crunch, splash, and creak. Several studies have shown strong associations between sound symbolic words, shapes (e.g., Bouba/Kiki effect), and materials. Here, we identified these material associations in Turkish sound symbolic words and then tested for their effect on softness perception. First, we used a rating task in a semantic differentiation method to extract the perceived softness dimensions from words and materials. We then tested whether Turkish onomatopoeic words can be used to manipulate the perceived softness of everyday materials such as honey, silk, or sand across different dimensions of softness. In the first preliminary study, we used 40 material videos and 29 adjectives in a rating task with a semantic differentiation method to extract the main softness dimensions. A principal component analysis revealed seven softness components, including Deformability, Viscosity, Surface Softness, and Granularity, in line with the literature. The second preliminary study used 27 onomatopoeic words and 21 adjectives in the same rating task. Again, the findings aligned with the literature, revealing dimensions such as Viscosity, Granularity, and Surface Softness. However, no factors related to Deformability were found due to the absence of sound symbolic words in this category. Next, we paired the onomatopoeic words and material videos based on their associations with each softness dimension. We conducted a new rating task, synchronously presenting material videos and spoken onomatopoeic words. We hypothesized that congruent word-video pairs would produce significantly higher ratings for dimension-related adjectives, while incongruent word-video pairs would decrease these ratings, and the ratings of unrelated adjectives would remain the same. Our results revealed that onomatopoeic words selectively alter the perceived material qualities, providing evidence and insight into the cross-modality of perceived softness.
Collapse
|
6
|
Cavdan M, Goktepe N, Drewing K, Doerschner K. Assessing the representational structure of softness activated by words. Sci Rep 2023; 13:8974. [PMID: 37268674 DOI: 10.1038/s41598-023-35169-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2022] [Accepted: 05/14/2023] [Indexed: 06/04/2023] Open
Abstract
Softness is an important material property that can be judged directly, by interacting with an object, but also indirectly, by simply looking at an image of a material. The latter is likely possible by filling in relevant multisensory information from prior experiences with soft materials. Such experiences are thought to lead to associations that make up our representations about perceptual softness. Here, we investigate the structure of this representational space when activated by words, and compare it to haptic and visual perceptual spaces that we obtained in earlier work. To this end, we performed an online study where people rated different sensory aspects of soft materials, presented as written names. We compared the results with the previous studies where identical ratings were made on the basis of visual and haptic information. Correlation and Procrustes analyses show that, overall, the representational spaces of verbally presented materials were similar to those obtained from haptic and visual experiments. However, a classifier analysis showed that verbal representations could better be predicted from those obtained from visual than from haptic experiments. In a second study we rule out that these larger discrepancies in representations between verbal and haptic conditions could be due to difficulties in material identification in haptic experiments. We discuss the results with respect to the recent idea that at perceived softness is a multidimensional construct.
Collapse
Affiliation(s)
- Müge Cavdan
- Experimental Psychology, Justus-Liebig-Universität Gießen, 35390, Gießen, Germany.
| | - Nedim Goktepe
- Philipps-University Marburg, 35037, Marburg, Germany
| | - Knut Drewing
- Experimental Psychology, Justus-Liebig-Universität Gießen, 35390, Gießen, Germany
| | - Katja Doerschner
- Experimental Psychology, Justus-Liebig-Universität Gießen, 35390, Gießen, Germany
- National Magnetic Resonance Research Center, Ankara, 06800, Turkey
| |
Collapse
|
7
|
Di Stefano N, Spence C. Roughness perception: A multisensory/crossmodal perspective. Atten Percept Psychophys 2022; 84:2087-2114. [PMID: 36028614 PMCID: PMC9481510 DOI: 10.3758/s13414-022-02550-y] [Citation(s) in RCA: 26] [Impact Index Per Article: 8.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 08/01/2022] [Indexed: 11/08/2022]
Abstract
Roughness is a perceptual attribute typically associated with certain stimuli that are presented in one of the spatial senses. In auditory research, the term is typically used to describe the harsh effects that are induced by particular sound qualities (i.e., dissonance) and human/animal vocalizations (e.g., screams, distress cries). In the tactile domain, roughness is a crucial factor determining the perceptual features of a surface. The same feature can also be ascertained visually, by means of the extraction of pattern features that determine the haptic quality of surfaces, such as grain size and density. By contrast, the term roughness has rarely been applied to the description of those stimuli perceived via the chemical senses. In this review, we take a critical look at the putative meaning(s) of the term roughness, when used in both unisensory and multisensory contexts, in an attempt to answer two key questions: (1) Is the use of the term 'roughness' the same in each modality when considered individually? and (2) Do crossmodal correspondences involving roughness match distinct perceptual features or (at least on certain occasions) do they merely pick-up on an amodal property? We start by examining the use of the term in the auditory domain. Next, we summarize the ways in which the term roughness has been used in the literature on tactile and visual perception, and in the domain of olfaction and gustation. Then, we move on to the crossmodal context, reviewing the literature on the perception of roughness in the audiovisual, audiotactile, and auditory-gustatory/olfactory domains. Finally, we highlight some limitations of the reviewed literature and we outline a number of key directions for future empirical research in roughness perception.
Collapse
Affiliation(s)
- Nicola Di Stefano
- National Research Council, Institute for Cognitive Sciences and Technologies, Rome, Italy.
| | | |
Collapse
|
8
|
|