1
|
Sáringer S, Kaposvári P, Benyhe A. Visual linguistic statistical learning is traceable through neural entrainment. Psychophysiology 2024; 61:e14575. [PMID: 38549442 DOI: 10.1111/psyp.14575] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2023] [Revised: 02/22/2024] [Accepted: 03/17/2024] [Indexed: 07/07/2024]
Abstract
The human brain can detect statistical regularities in the environment across a wide variety of contexts. The importance of this process is well-established not just in language acquisition but across different modalities; in addition, several neural correlates of statistical learning have been identified. A current technique for tracking the emergence of regularity learning and localizing its neural background is frequency tagging (FT). FT can detect neural entrainment not only to the frequency of stimulus presentation but also to that of a hidden structure. Auditory learning paradigms with linguistic and nonlinguistic stimuli, along with a visual paradigm using nonlinguistic stimuli, have already been tested with FT. To complete the picture, we conducted an FT experiment using written syllables as stimuli and a hidden triplet structure. Both behavioral and neural entrainment data showed evidence of structure learning. In addition, we localized two electrode clusters related to the process, which spread across the frontal and parieto-occipital areas, similar to previous findings. Accordingly, we conclude that fast-paced visual linguistic regularities can be acquired and are traceable through neural entrainment. In comparison with the literature, our findings support the view that statistical learning involves a domain-general network.
Collapse
Affiliation(s)
- Szabolcs Sáringer
- Department of Physiology, Albert Szent-Györgyi Medical School, University of Szeged, Szeged, Hungary
| | - Péter Kaposvári
- Department of Physiology, Albert Szent-Györgyi Medical School, University of Szeged, Szeged, Hungary
| | - András Benyhe
- Department of Physiology, Albert Szent-Györgyi Medical School, University of Szeged, Szeged, Hungary
| |
Collapse
|
2
|
Lee SMK, Law NSH, Tong SX. Unraveling Temporal Dynamics of Multidimensional Statistical Learning in Implicit and Explicit Systems: An X-Way Hypothesis. Cogn Sci 2024; 48:e13437. [PMID: 38564270 DOI: 10.1111/cogs.13437] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2023] [Revised: 03/05/2024] [Accepted: 03/13/2024] [Indexed: 04/04/2024]
Abstract
Statistical learning enables humans to involuntarily process and utilize different kinds of patterns from the environment. However, the cognitive mechanisms underlying the simultaneous acquisition of multiple regularities from different perceptual modalities remain unclear. A novel multidimensional serial reaction time task was developed to test 40 participants' ability to learn simple first-order and complex second-order relations between uni-modal visual and cross-modal audio-visual stimuli. Using the difference in reaction times between sequenced and random stimuli as the index of domain-general statistical learning, a significant difference and dissociation of learning occurred between the initial and final learning phases. Furthermore, we used a negative and positive occurrence-frequency-and-reaction-time correlation to indicate implicit and explicit learning, respectively, and found that learning simple uni-modal patterns involved an implicit-to-explicit segue, while acquiring complex cross-modal patterns required an explicit-to-implicit segue, resulting in a X-shape crossing of regularity learning. Thus, we propose an X-way hypothesis to elucidate the dynamic interplay between the implicit and explicit systems at two distinct stages when acquiring various regularities in a multidimensional probability space.
Collapse
Affiliation(s)
- Stephen Man-Kit Lee
- Academic Unit of Human Communication, Learning, and Development, The University of Hong Kong
| | - Nicole Sin Hang Law
- Academic Unit of Human Communication, Learning, and Development, The University of Hong Kong
- Department of Education, University of Oxford
| | - Shelley Xiuli Tong
- Academic Unit of Human Communication, Learning, and Development, The University of Hong Kong
| |
Collapse
|
3
|
Pasqualotto A, Cochrane A, Bavelier D, Altarelli I. A novel task and methods to evaluate inter-individual variation in audio-visual associative learning. Cognition 2024; 242:105658. [PMID: 37952371 DOI: 10.1016/j.cognition.2023.105658] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2022] [Revised: 10/24/2023] [Accepted: 10/31/2023] [Indexed: 11/14/2023]
Abstract
Learning audio-visual associations is foundational to a number of real-world skills, such as reading acquisition or social communication. Characterizing individual differences in such learning has therefore been of interest to researchers in the field. Here, we present a novel audio-visual associative learning task designed to efficiently capture inter-individual differences in learning, with the added feature of using non-linguistic stimuli, so as to unconfound language and reading proficiency of the learner from their more domain-general learning capability. By fitting trial-by-trial performance in our novel learning task using simple-to-use statistical tools, we demonstrate the expected inter-individual variability in learning rate as well as high precision in its estimation. We further demonstrate that such measured learning rate is linked to working memory performance in Italian-speaking (N = 58) and French-speaking (N = 51) adults. Finally, we investigate the extent to which learning rate in our task, which measures cross-modal audio-visual associations while mitigating familiarity confounds, predicts reading ability across participants with different linguistic backgrounds. The present work thus introduces a novel non-linguistic audio-visual associative learning task that can be used across languages. In doing so, it brings a new tool to researchers in the various domains that rely on multi-sensory integration from reading to social cognition or socio-emotional learning.
Collapse
Affiliation(s)
- Angela Pasqualotto
- Faculty of Psychology and Education Sciences (FPSE), University of Geneva, Geneva, Switzerland; Campus Biotech, Geneva, Switzerland
| | - Aaron Cochrane
- Faculty of Psychology and Education Sciences (FPSE), University of Geneva, Geneva, Switzerland; Campus Biotech, Geneva, Switzerland
| | - Daphne Bavelier
- Faculty of Psychology and Education Sciences (FPSE), University of Geneva, Geneva, Switzerland; Campus Biotech, Geneva, Switzerland.
| | | |
Collapse
|
4
|
Sweet SJ, Van Hedger SC, Batterink LJ. Of words and whistles: Statistical learning operates similarly for identical sounds perceived as speech and non-speech. Cognition 2024; 242:105649. [PMID: 37871411 DOI: 10.1016/j.cognition.2023.105649] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2023] [Revised: 10/11/2023] [Accepted: 10/13/2023] [Indexed: 10/25/2023]
Abstract
Statistical learning is an ability that allows individuals to effortlessly extract patterns from the environment, such as sound patterns in speech. Some prior evidence suggests that statistical learning operates more robustly for speech compared to non-speech stimuli, supporting the idea that humans are predisposed to learn language. However, any apparent statistical learning advantage for speech could be driven by signal acoustics, rather than the subjective perception per se of sounds as speech. To resolve this issue, the current study assessed whether there is a statistical learning advantage for ambiguous sounds that are subjectively perceived as speech-like compared to the same sounds perceived as non-speech, thereby controlling for acoustic features. We first induced participants to perceive sine-wave speech (SWS)-a degraded form of speech not immediately perceptible as speech-as either speech or non-speech. After this induction phase, participants were exposed to a continuous stream of repeating trisyllabic nonsense words, composed of SWS syllables, and then completed an explicit familiarity rating task and an implicit target detection task to assess learning. Critically, participants showed robust and equivalent performance on both measures, regardless of their subjective speech perception. In contrast, participants who perceived the SWS syllables as more speech-like showed better detection of individual syllables embedded in speech streams. These results suggest that speech perception facilitates processing of individual sounds, but not the ability to extract patterns across sounds. Our findings suggest that statistical learning is not influenced by the perceived linguistic relevance of sounds, and that it may be conceptualized largely as an automatic, stimulus-driven mechanism.
Collapse
Affiliation(s)
- Sierra J Sweet
- Department of Psychology, Western University, London, ON, Canada.
| | - Stephen C Van Hedger
- Department of Psychology, Western University, London, ON, Canada; Western Institute for Neuroscience, Western University, London, ON, Canada; Department of Psychology, Huron University College, London, ON, Canada.
| | - Laura J Batterink
- Department of Psychology, Western University, London, ON, Canada; Western Institute for Neuroscience, Western University, London, ON, Canada.
| |
Collapse
|
5
|
Sáringer S, Fehér Á, Sáry G, Kaposvári P. Gamma oscillations in visual statistical learning correlate with individual behavioral differences. Front Behav Neurosci 2023; 17:1285773. [PMID: 38025386 PMCID: PMC10663268 DOI: 10.3389/fnbeh.2023.1285773] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Accepted: 10/23/2023] [Indexed: 12/01/2023] Open
Abstract
Statistical learning is assumed to be a fundamentally general sensory process across modalities, age, other cognitive functions, and even species. Despite this general role, behavioral testing on regularity acquisition shows great variance among individuals. The current study aimed to find neural correlates of visual statistical learning showing a correlation with behavioral results. Based on a pilot study, we conducted an EEG study where participants were exposed to associated stimulus pairs; the acquisition was tested through a familiarity test. We identified an oscillation in the gamma range (40-70 Hz, 0.5-0.75 s post-stimulus), which showed a positive correlation with the behavioral results. This change in activity was located in a left frontoparietal cluster. Based on its latency and location, this difference was identified as a late gamma activity, a correlate of model-based learning. Such learning is a summary of several top-down mechanisms that modulate the recollection of statistical relationships such as the capacity of working memory or attention. These results suggest that, during acquisition, individual behavioral variance is influenced by dominant learning processes which affect the recall of previously gained information.
Collapse
Affiliation(s)
| | | | | | - Péter Kaposvári
- Department of Physiology, Albert Szent-Gyögyi Medical School, University of Szeged, Szeged, Hungary
| |
Collapse
|
6
|
Wilts S, Haider H. Concurrent visual sequence learning. PSYCHOLOGICAL RESEARCH 2023; 87:2086-2100. [PMID: 36947194 PMCID: PMC10457409 DOI: 10.1007/s00426-023-01810-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2022] [Accepted: 02/15/2023] [Indexed: 03/23/2023]
Abstract
Many researchers in the field of implicit statistical learning agree that there does not exist one general implicit learning mechanism, but rather, that implicit learning takes place in highly specialized encapsulated modules. However, the exact representational content of these modules is still under debate. While there is ample evidence for a distinction between modalities (e.g., visual, auditory perception), the representational content of the modules might even be distinguished by features within the same modalities (e.g., location, color, and shape within the visual modality). In implicit sequence learning, there is evidence for the latter hypothesis, as a stimulus-color sequence can be learned concurrently with a stimulus-location sequence. Our aim was to test whether this also holds true for non-spatial features within the visual modality. This has been shown in artificial grammar learning, but not yet in implicit sequence learning. Hence, in Experiment 1, we replicated an artificial grammar learning experiment of Conway and Christiansen (2006) in which participants were supposed to learn color and shape grammars concurrently. In Experiment 2, we investigated concurrent learning of sequences with an implicit sequence learning paradigm: the serial reaction time task. Here, we found evidence for concurrent learning of two sequences, a color and shape sequence. Overall, the findings converge to the assumption that implicit learning might be based on features.
Collapse
Affiliation(s)
- Sarah Wilts
- Department of Psychology, University of Cologne, Richard-Strauss-Str. 2, 50931 Cologne, Germany
| | - Hilde Haider
- Department of Psychology, University of Cologne, Richard-Strauss-Str. 2, 50931 Cologne, Germany
| |
Collapse
|
7
|
Murray CA, Shams L. Crossmodal interactions in human learning and memory. Front Hum Neurosci 2023; 17:1181760. [PMID: 37266327 PMCID: PMC10229776 DOI: 10.3389/fnhum.2023.1181760] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2023] [Accepted: 05/02/2023] [Indexed: 06/03/2023] Open
Abstract
Most studies of memory and perceptual learning in humans have employed unisensory settings to simplify the study paradigm. However, in daily life we are often surrounded by complex and cluttered scenes made up of many objects and sources of sensory stimulation. Our experiences are, therefore, highly multisensory both when passively observing the world and when acting and navigating. We argue that human learning and memory systems are evolved to operate under these multisensory and dynamic conditions. The nervous system exploits the rich array of sensory inputs in this process, is sensitive to the relationship between the sensory inputs, and continuously updates sensory representations, and encodes memory traces based on the relationship between the senses. We review some recent findings that demonstrate a range of human learning and memory phenomena in which the interactions between visual and auditory modalities play an important role, and suggest possible neural mechanisms that can underlie some surprising recent findings. We outline open questions as well as directions of future research to unravel human perceptual learning and memory.
Collapse
Affiliation(s)
- Carolyn A. Murray
- Department of Psychology, University of California, Los Angeles, Los Angeles, CA, United States
| | - Ladan Shams
- Department of Psychology, University of California, Los Angeles, Los Angeles, CA, United States
- Department of Bioengineering, Neuroscience Interdepartmental Program, University of California, Los Angeles, Los Angeles, CA, United States
| |
Collapse
|
8
|
Polgári P, Jovanovic L, van Wassenhove V, Giersch A. The processing of subthreshold visual temporal order is transitory and motivation-dependent. Sci Rep 2023; 13:7699. [PMID: 37169810 PMCID: PMC10175535 DOI: 10.1038/s41598-023-34392-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2022] [Accepted: 04/28/2023] [Indexed: 05/13/2023] Open
Abstract
Processing a sequence of events is different from encoding the relative order of the elements composing the sequence. Whether order processing arises automatically from the sequential processing of events is yet unknown, however the literature suggests that order processing can occur at an automatic level when the order of stimuli is not detected consciously. In the present study, we aimed to investigate the question of automatic order processing in a difficult visual task where participants identified one among two possible target luminances. The luminance of the targets was contingent on the order of presentation of two visual cues separated by a subthreshold asynchrony. Participants' performance was compared to that in a control condition where the cues were presented synchronously. In a first experiment, participants' performance benefited from the use of subthreshold order information compared to the control condition, however this facilitation effect was transient and disappeared over the course of the experiment. In a second experiment, we investigated and confirmed the role of motivation, via a monetary incentive, on the previously observed effect. Taken together, our results suggest that the processing of temporal order of sub-threshold asynchronies is possible, although fragile and likely dependent on task requirements.
Collapse
Affiliation(s)
- Patrik Polgári
- University of Strasbourg, INSERM U1114, Strasbourg, France
| | - Ljubica Jovanovic
- University of Strasbourg, INSERM U1114, Strasbourg, France
- Laboratoire des Systèmes Perceptifs, École Normale Supérieure, PSL University & CNRS, Paris, France
| | - Virginie van Wassenhove
- Cognitive Neuroimaging Unit, CEA, INSERM, CNRS, Neurospin, Université Paris-Saclay, 91191, Gif/Yvette, France
| | - Anne Giersch
- University of Strasbourg, INSERM U1114, Strasbourg, France.
- Department of Psychiatry, University Hospital of Strasbourg, 1, Pl de L'Hôpital, 67000, Strasbourg, France.
| |
Collapse
|
9
|
MacIntyre AD, Lo HYJ, Cross I, Scott S. Task-irrelevant auditory metre shapes visuomotor sequential learning. PSYCHOLOGICAL RESEARCH 2023; 87:872-893. [PMID: 35690927 PMCID: PMC10017598 DOI: 10.1007/s00426-022-01690-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2021] [Accepted: 05/17/2022] [Indexed: 11/24/2022]
Abstract
The ability to learn and reproduce sequences is fundamental to every-day life, and deficits in sequential learning are associated with developmental disorders such as specific language impairment. Individual differences in sequential learning are usually investigated using the serial reaction time task (SRTT), wherein a participant responds to a series of regularly timed, seemingly random visual cues that in fact follow a repeating deterministic structure. Although manipulating inter-cue interval timing has been shown to adversely affect sequential learning, the role of metre (the patterning of salience across time) remains unexplored within the regularly timed, visual SRTT. The current experiment consists of an SRTT adapted to include task-irrelevant auditory rhythms conferring a sense of metre. We predicted that (1) participants' (n = 41) reaction times would reflect the auditory metric structure; (2) that disrupting the correspondence between the learned visual sequence and auditory metre would impede performance; and (3) that individual differences in sensitivity to rhythm would predict the magnitude of these effects. Altering the relationship via a phase shift between the trained visual sequence and auditory metre slowed reaction times. Sensitivity to rhythm was predictive of reaction times over all. In an exploratory analysis, we, moreover, found that approximately half of participants made systematically different responses to visual cues on the basis of the cues' position within the auditory metre. We demonstrate the influence of auditory temporal structures on visuomotor sequential learning in a widely used task where metre and timing are rarely considered. The current results indicate sensitivity to metre as a possible latent factor underpinning individual differences in SRTT performance.
Collapse
Affiliation(s)
- Alexis Deighton MacIntyre
- Institute of Cognitive Neuroscience, University College London, London, UK. .,Centre for Music and Science, University of Cambridge, Cambridge, UK. .,MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK.
| | | | - Ian Cross
- Centre for Music and Science, University of Cambridge, Cambridge, UK
| | - Sophie Scott
- Institute of Cognitive Neuroscience, University College London, London, UK
| |
Collapse
|
10
|
Smyre SA, Bean NL, Stein BE, Rowland BA. Predictability alters multisensory responses by modulating unisensory inputs. Front Neurosci 2023; 17:1150168. [PMID: 37065927 PMCID: PMC10090419 DOI: 10.3389/fnins.2023.1150168] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2023] [Accepted: 03/13/2023] [Indexed: 03/30/2023] Open
Abstract
The multisensory (deep) layers of the superior colliculus (SC) play an important role in detecting, localizing, and guiding orientation responses to salient events in the environment. Essential to this role is the ability of SC neurons to enhance their responses to events detected by more than one sensory modality and to become desensitized (‘attenuated’ or ‘habituated’) or sensitized (‘potentiated’) to events that are predictable via modulatory dynamics. To identify the nature of these modulatory dynamics, we examined how the repetition of different sensory stimuli affected the unisensory and multisensory responses of neurons in the cat SC. Neurons were presented with 2HZ stimulus trains of three identical visual, auditory, or combined visual–auditory stimuli, followed by a fourth stimulus that was either the same or different (‘switch’). Modulatory dynamics proved to be sensory-specific: they did not transfer when the stimulus switched to another modality. However, they did transfer when switching from the visual–auditory stimulus train to either of its modality-specific component stimuli and vice versa. These observations suggest that predictions, in the form of modulatory dynamics induced by stimulus repetition, are independently sourced from and applied to the modality-specific inputs to the multisensory neuron. This falsifies several plausible mechanisms for these modulatory dynamics: they neither produce general changes in the neuron’s transform, nor are they dependent on the neuron’s output.
Collapse
|
11
|
Quintero SI, Shams L, Kamal K. Changing the Tendency to Integrate the Senses. Brain Sci 2022; 12:brainsci12101384. [PMID: 36291318 PMCID: PMC9599885 DOI: 10.3390/brainsci12101384] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2022] [Revised: 10/04/2022] [Accepted: 10/05/2022] [Indexed: 11/16/2022] Open
Abstract
Integration of sensory signals that emanate from the same source, such as the visual of lip articulations and the sound of the voice of a speaking individual, can improve perception of the source signal (e.g., speech). Because momentary sensory inputs are typically corrupted with internal and external noise, there is almost always a discrepancy between the inputs, facing the perceptual system with the problem of determining whether the two signals were caused by the same source or different sources. Thus, whether or not multisensory stimuli are integrated and the degree to which they are bound is influenced by factors such as the prior expectation of a common source. We refer to this factor as the tendency to bind stimuli, or for short, binding tendency. In theory, the tendency to bind sensory stimuli can be learned by experience through the acquisition of the probabilities of the co-occurrence of the stimuli. It can also be influenced by cognitive knowledge of the environment. The binding tendency varies across individuals and can also vary within an individual over time. Here, we review the studies that have investigated the plasticity of binding tendency. We discuss the protocols that have been reported to produce changes in binding tendency, the candidate learning mechanisms involved in this process, the possible neural correlates of binding tendency, and outstanding questions pertaining to binding tendency and its plasticity. We conclude by proposing directions for future research and argue that understanding mechanisms and recipes for increasing binding tendency can have important clinical and translational applications for populations or individuals with a deficiency in multisensory integration.
Collapse
Affiliation(s)
- Saul I Quintero
- Department of Psychology, University of California, Los Angeles, CA 90095, USA
| | - Ladan Shams
- Department of Psychology, University of California, Los Angeles, CA 90095, USA
- Department of Bioengineering, University of California, Los Angeles, CA 90089, USA
- Neuroscience Interdepartmental Program, University of California, Los Angeles, CA 90089, USA
| | - Kimia Kamal
- Department of Psychology, University of California, Los Angeles, CA 90095, USA
| |
Collapse
|
12
|
Online measurement of learning temporal statistical structure in categorization tasks. Mem Cognit 2022; 50:1530-1545. [PMID: 35377057 PMCID: PMC9508059 DOI: 10.3758/s13421-022-01302-5] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 03/10/2022] [Indexed: 11/08/2022]
Abstract
AbstractThe ability to grasp relevant patterns from a continuous stream of environmental information is called statistical learning. Although the representations that emerge during visual statistical learning (VSL) are well characterized, little is known about how they are formed. We developed a sensitive behavioral design to characterize the VSL trajectory during ongoing task performance. In sequential categorization tasks, we assessed two previously identified VSL markers: priming of the second predictable image in a pair manifested by a reduced reaction time (RT) and greater accuracy, and the anticipatory effect on the first image revealed by a longer RT. First, in Experiment 1A, we used an adapted paradigm and replicated these VSL markers; however, they appeared to be confounded by motor learning. Next, in Experiment 1B, we confirmed the confounding influence of motor learning. To assess VSL without motor learning, in Experiment 2 we (1) simplified the categorization task, (2) raised the number of subjects and image repetitions, and (3) increased the number of single unpaired images. Using linear mixed-effect modeling and estimated marginal means of linear trends, we found that the RT curves differed significantly between predictable paired and control single images. Further, the VSL curve fitted a logarithmic model, suggesting a rapid learning process. These results suggest that our paradigm in Experiment 2 seems to be a viable online tool to monitor the behavioral correlates of unsupervised implicit VSL.
Collapse
|
13
|
Li J, Deng SW. Facilitation and interference effects of the multisensory context on learning: a systematic review and meta-analysis. PSYCHOLOGICAL RESEARCH 2022; 87:1334-1352. [DOI: 10.1007/s00426-022-01733-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2022] [Accepted: 08/28/2022] [Indexed: 11/24/2022]
|
14
|
Agrawal T, Schachner A. Hearing water temperature: characterizing the development of nuanced perception of sound sources. Dev Sci 2022; 26:e13321. [PMID: 36068928 DOI: 10.1111/desc.13321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2022] [Revised: 08/18/2022] [Accepted: 08/22/2022] [Indexed: 11/30/2022]
Abstract
Without conscious thought, listeners link events in the world to sounds they hear. We study one surprising example: Adults can judge the temperature of water simply from hearing it being poured. We test development of the ability to hear water temperature, with the goal of informing developmental theories regarding the origins and cognitive bases of nuanced sound source judgments. We first confirmed that adults accurately distinguished the sounds of hot and cold water (pre-registered Exps. 1, 2; total N = 384), even though many were unaware or uncertain of this ability. By contrast, children showed protracted development of this skill over the course of middle childhood (Exps. 2, 3; total N = 178). In spite of accurately identifying other sounds and hot/cold images, older children (7-11 years) but not younger children (3-6 years) reliably distinguished the sounds of hot and cold water. Accuracy increased with age; 11 year old's performance was similar to adults'. Adults also showed individual differences in accuracy that were predicted by their amount of prior relevant experience (Exp. 1). Experience may similarly play a role in children's performance; differences in auditory sensitivity and multimodal integration may also contribute to young children's failures. The ability to hear water temperature develops slowly over childhood, such that nuanced auditory information that is easily and quickly accessible to adults is not available to guide young children's behavior. Adults can make nuanced judgments from sound, including accurately judging the temperature of water from the sound of it being poured. Children showed protracted development of this skill over the course of middle childhood, such that 7-11 year-olds reliably succeeded while 3-6 year-olds performed at chance. Developmental changes may be due to experience (adults with greater relevant experience showed higher accuracy) and development of multimodal integration and auditory sensitivity. Young children may not detect subtle auditory information that adults easily perceive. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
| | - Adena Schachner
- Department of Psychology, University of California, San Diego, USA
| |
Collapse
|
15
|
Chalas N, Karagiorgis A, Bamidis P, Paraskevopoulos E. The impact of musical training in symbolic and non-symbolic audiovisual judgements of magnitude. PLoS One 2022; 17:e0266165. [PMID: 35511806 PMCID: PMC9070945 DOI: 10.1371/journal.pone.0266165] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2021] [Accepted: 03/15/2022] [Indexed: 11/30/2022] Open
Abstract
Quantity estimation can be represented in either an analog or symbolic manner and recent evidence now suggests that analog and symbolic representation of quantities interact. Nonetheless, those two representational forms of quantities may be enhanced by convergent multisensory information. Here, we elucidate those interactions using high-density electroencephalography (EEG) and an audiovisual oddball paradigm. Participants were presented simultaneous audiovisual tokens in which the co-varying pitch of tones was combined with the embedded cardinality of dot patterns. Incongruencies were elicited independently from symbolic and non-symbolic modality within the audio-visual percept, violating the newly acquired rule that “the higher the pitch of the tone, the larger the cardinality of the figure.” The effect of neural plasticity in symbolic and non-symbolic numerical representations of quantities was investigated through a cross-sectional design, comparing musicians to musically naïve controls. Individual’s cortical activity was reconstructed and statistically modeled for a predefined time-window of the evoked response (130–170 ms). To summarize, we show that symbolic and non-symbolic processing of magnitudes is re-organized in cortical space, with professional musicians showing altered activity in motor and temporal areas. Thus, we argue that the symbolic representation of quantities is altered through musical training.
Collapse
Affiliation(s)
- Nikos Chalas
- Institute for Biomagnetism and Biosignal analysis, University of Münster, Münster, Germany
- School of Medicine, Faculty of Health Sciences, Aristotle University of Thessaloniki, Thessaloniki, Greece
| | - Alexandros Karagiorgis
- School of Medicine, Faculty of Health Sciences, Aristotle University of Thessaloniki, Thessaloniki, Greece
| | - Panagiotis Bamidis
- School of Medicine, Faculty of Health Sciences, Aristotle University of Thessaloniki, Thessaloniki, Greece
| | - Evangelos Paraskevopoulos
- School of Medicine, Faculty of Health Sciences, Aristotle University of Thessaloniki, Thessaloniki, Greece
- Department of Psychology, University of Cyprus, Nicosia, Cyprus
- * E-mail:
| |
Collapse
|
16
|
Lagarrigue Y, Cappe C, Tallet J. Regular rhythmic and audio-visual stimulations enhance procedural learning of a perceptual-motor sequence in healthy adults: A pilot study. PLoS One 2021; 16:e0259081. [PMID: 34780497 PMCID: PMC8592429 DOI: 10.1371/journal.pone.0259081] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/11/2021] [Accepted: 10/12/2021] [Indexed: 12/02/2022] Open
Abstract
Procedural learning is essential for the effortless execution of many everyday life activities. However, little is known about the conditions influencing the acquisition of procedural skills. The literature suggests that sensory environment may influence the acquisition of perceptual-motor sequences, as tested by a Serial Reaction Time Task. In the current study, we investigated the effects of auditory stimulations on procedural learning of a visuo-motor sequence. Given that the literature shows that regular rhythmic auditory rhythm and multisensory stimulations improve motor speed, we expected to improve procedural learning (reaction times and errors) with repeated practice with auditory stimulations presented either simultaneously with visual stimulations or with a regular tempo, compared to control conditions (e.g., with irregular tempo). Our results suggest that both congruent audio-visual stimulations and regular rhythmic auditory stimulations promote procedural perceptual-motor learning. On the contrary, auditory stimulations with irregular or very quick tempo alter learning. We discuss how regular rhythmic multisensory stimulations may improve procedural learning with respect of a multisensory rhythmic integration process.
Collapse
Affiliation(s)
- Yannick Lagarrigue
- ToNIC, Toulouse NeuroImaging Center, Université de Toulouse, Inserm, UPS, Toulouse, France
- * E-mail:
| | - Céline Cappe
- Cerco, Centre de Recherche Cerveau et Cognition, Université de Toulouse, CNRS, UMR 5549, Toulouse, France
| | - Jessica Tallet
- ToNIC, Toulouse NeuroImaging Center, Université de Toulouse, Inserm, UPS, Toulouse, France
| |
Collapse
|
17
|
An Overview of Olfactory Displays in Education and Training. MULTIMODAL TECHNOLOGIES AND INTERACTION 2021. [DOI: 10.3390/mti5100064] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
This paper describes an overview of olfactory displays (human–computer interfaces that generate and diffuse an odor to a user to stimulate their sense of smell) that have been proposed and researched for supporting education and training. Past research has shown that olfaction (the sense of smell) can support memorization of information, stimulate information recall, and help immerse learners and trainees into educational virtual environments, as well as complement and/or supplement other human sensory channels for learning. This paper begins with an introduction to olfaction and olfactory displays, and a review of techniques for storing, generating and diffusing odors at the computer interface. The paper proceeds with a discussion on educational theories that support olfactory displays for education and training, and a literature review on olfactory displays that support learning and training. Finally, the paper summarizes the advantages and challenges regarding the development and application of olfactory displays for education and training.
Collapse
|
18
|
Nazaré CJ, Oliveira AM. Effects of Audiovisual Presentations on Visual Localization Errors: One or Several Multisensory Mechanisms? Multisens Res 2021; 34:1-35. [PMID: 33882452 DOI: 10.1163/22134808-bja10048] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2020] [Accepted: 03/30/2021] [Indexed: 11/19/2022]
Abstract
The present study examines the extent to which temporal and spatial properties of sound modulate visual motion processing in spatial localization tasks. Participants were asked to locate the place at which a moving visual target unexpectedly vanished. Across different tasks, accompanying sounds were factorially varied within subjects as to their onset and offset times and/or positions relative to visual motion. Sound onset had no effect on the localization error. Sound offset was shown to modulate the perceived visual offset location, both for temporal and spatial disparities. This modulation did not conform to attraction toward the timing or location of the sounds but, demonstrably in the case of temporal disparities, to bimodal enhancement instead. Favorable indications to a contextual effect of audiovisual presentations on interspersed visual-only trials were also found. The short sound-leading offset asynchrony had equivalent benefits to audiovisual offset synchrony, suggestive of the involvement of early-level mechanisms, constrained by a temporal window, at these conditions. Yet, we tentatively hypothesize that the whole of the results and how they compare with previous studies requires the contribution of additional mechanisms, including learning-detection of auditory-visual associations and cross-sensory spread of endogenous attention.
Collapse
Affiliation(s)
- Cristina Jordão Nazaré
- Instituto Politécnico de Coimbra, ESTESC - Coimbra Health School, Audiologia, Coimbra, Portugal
| | | |
Collapse
|
19
|
Röttger E, Zhao F, Gaschler R, Haider H. Why Does Dual-Tasking Hamper Implicit Sequence Learning? J Cogn 2021; 4:1. [PMID: 33506167 PMCID: PMC7792471 DOI: 10.5334/joc.136] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2020] [Accepted: 10/15/2020] [Indexed: 11/20/2022] Open
Abstract
Research on the limitations of dual-tasking might profit from using setups with a predictable sequence of stimuli and responses and assessing the acquisition of this sequence. Detrimental effects of dual-tasking on implicit sequence learning in the serial reaction time task (SRTT; Nissen & Bullemer, 1987) - when paired with an uncorrelated task - have been attributed to participants' lack of separating the streams of events in either task. Assuming that co-occurring events are automatically integrated, we reasoned that participants could need to first learn which events co-occur, before they can acquire sequence knowledge. In the training phase, we paired an 8-element visual-manual SRTT with an auditory-vocal task. Afterwards, we tested under single-tasking conditions whether SRTT sequence knowledge had been acquired. By applying different variants of probabilistic SRTT-tone pairings across three experiments, we tested what type of predictive relationship was needed to preserve sequence learning. In Experiment 1, where half of the SRTT-elements were paired to 100% with one specific tone and the other half randomly, only the fixedly paired elements were learned. Yet, no sequence learning was found when each of the eight SRTT-elements was paired with tone identity in a 75%-25% ratio (Experiment 2). Sequence learning was, however, intact when the 75%-25% ratio was applied to the four SRTT target locations instead (Experiment 3). The results suggest that participants (when lacking a separation of the task representations while dual-tasking) can learn a sequence inherent in one of two tasks to the extent that across-task contingencies can be learned first.
Collapse
Affiliation(s)
- Eva Röttger
- Department of Psychology, University of Bremen, Hochschulring 18, 28359 Bremen, DE
| | - Fang Zhao
- Department of Psychology, FernUniversität in Hagen, Universitätsstr. 33, 58084 Hagen, DE
| | - Robert Gaschler
- Department of Psychology, FernUniversität in Hagen, Universitätsstr. 33, 58084 Hagen, DE
| | - Hilde Haider
- Department of Psychology, University of Cologne, Richard-Strauss-Str. 2, 50931 Köln, DE
| |
Collapse
|
20
|
Conway CM. How does the brain learn environmental structure? Ten core principles for understanding the neurocognitive mechanisms of statistical learning. Neurosci Biobehav Rev 2020; 112:279-299. [PMID: 32018038 PMCID: PMC7211144 DOI: 10.1016/j.neubiorev.2020.01.032] [Citation(s) in RCA: 98] [Impact Index Per Article: 24.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2019] [Revised: 01/22/2020] [Accepted: 01/25/2020] [Indexed: 10/25/2022]
Abstract
Despite a growing body of research devoted to the study of how humans encode environmental patterns, there is still no clear consensus about the nature of the neurocognitive mechanisms underpinning statistical learning nor what factors constrain or promote its emergence across individuals, species, and learning situations. Based on a review of research examining the roles of input modality and domain, input structure and complexity, attention, neuroanatomical bases, ontogeny, and phylogeny, ten core principles are proposed. Specifically, there exist two sets of neurocognitive mechanisms underlying statistical learning. First, a "suite" of associative-based, automatic, modality-specific learning mechanisms are mediated by the general principle of cortical plasticity, which results in improved processing and perceptual facilitation of encountered stimuli. Second, an attention-dependent system, mediated by the prefrontal cortex and related attentional and working memory networks, can modulate or gate learning and is necessary in order to learn nonadjacent dependencies and to integrate global patterns across time. This theoretical framework helps clarify conflicting research findings and provides the basis for future empirical and theoretical endeavors.
Collapse
Affiliation(s)
- Christopher M Conway
- Center for Childhood Deafness, Language, and Learning, Boys Town National Research Hospital, Omaha, NE, United States.
| |
Collapse
|
21
|
Artacho MÁ, Alcántara E, Martínez N. Multisensory Analysis of Consumer-Product Interaction During Ceramic Tile Shopping Experiences. Multisens Res 2020; 33:213-249. [PMID: 31648188 DOI: 10.1163/22134808-20191391] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2018] [Accepted: 09/27/2019] [Indexed: 11/19/2022]
Abstract
The need to design products that engage several senses has being increasingly recognised by design and marketing professionals. Many works analyse the impact of sensory stimuli on the hedonic, cognitive, and emotional responses of consumers, as well as on their satisfaction and intention to purchase. However, there is much less information about the utilitarian dimension related to a sensory non-reflective analysis of the tangible elements of the experience, the sequential role played by different senses, and their relative importance. This work analyses the sensorial dimension of consumer interactions in shops. Consumers were filmed in two ceramic tile shops and their behaviour was analysed according to a previously validated checklist. Sequence of actions, their frequency of occurrence, and the duration of inspections were recorded, and consumers were classified according to their sensory exploration strategies. Results show that inspection patterns are intentional but shifting throughout the interaction. Considering the whole sequence, vision is the dominant sense followed by touch. However, sensory dominance varies throughout the sequence. The dominance differences appear between all senses and within the senses of vision, touch and audition. Cluster analysis classified consumers into two groups, those who were more interactive and those who were visual and passive evaluators. These results are very important for understanding consumer interaction patterns, which senses are involved (including their importance and hierarchy), and which sensory properties of tiles are evaluated during the shopping experience. Moreover, this information is crucial for setting design guidelines to improve sensory interactions and bridge sensory demands with product features.
Collapse
Affiliation(s)
- Miguel Ángel Artacho
- 1Department of Engineering Projects, Universitat Politècnica de València, Building 5J, Camino de Vera s/n. 46022 Valencia, Spain
| | - Enrique Alcántara
- 2Institute of Biomechanics of Valencia, Universitat Politècnica de València, Building 9C, Camino de Vera s/n. 46022 Valencia, Spain
| | | |
Collapse
|
22
|
Audiohaptic Feedback Enhances Motor Performance in a Low-Fidelity Simulated Drilling Task. Brain Sci 2019; 10:brainsci10010021. [PMID: 31906192 PMCID: PMC7016775 DOI: 10.3390/brainsci10010021] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/26/2019] [Revised: 12/27/2019] [Accepted: 12/29/2019] [Indexed: 11/17/2022] Open
Abstract
When used in educational settings, simulations utilizing virtual reality (VR) technologies can reduce training costs while providing a safe and effective learning environment. Tasks can be easily modified to maximize learning objectives of different levels of trainees (e.g., novice, intermediate, expert), and can be repeated for the development of psychomotor skills. VR offers a multisensory experience, providing visual, auditory, and haptic sensations with varying levels of fidelity. While simulating visual and auditory stimuli is relatively easy and cost-effective, similar representations of haptic sensation still require further development. Evidence suggests that mixing high- and low-fidelity realistic sensations (e.g., audition and haptic) can improve the overall perception of realism, however, whether this also leads to improved performance has not been examined. The current study examined whether audiohaptic stimuli presented in a virtual drilling task can lead to improved motor performance and subjective realism, compared to auditory stimuli alone. Right-handed participants (n = 16) completed 100 drilling trials of each stimulus type. Performance measures indicated that participants overshot the target during auditory trials, and undershot the target during audiohaptic trials. Undershooting is thought to be indicative of improved performance, optimizing both time and energy requirements.
Collapse
|
23
|
Altarelli I, Dehaene-Lambertz G, Bavelier D. Individual differences in the acquisition of non-linguistic audio-visual associations in 5 year olds. Dev Sci 2019; 23:e12913. [PMID: 31608547 DOI: 10.1111/desc.12913] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2018] [Revised: 08/02/2019] [Accepted: 09/27/2019] [Indexed: 11/29/2022]
Abstract
Audio-visual associative learning - at least when linguistic stimuli are employed - is known to rely on core linguistic skills such as phonological awareness. Here we ask whether this would also be the case in a task that does not manipulate linguistic information. Another question of interest is whether executive skills, often found to support learning, may play a larger role in a non-linguistic audio-visual associative task compared to a linguistic one. We present a new task that measures learning when having to associate non-linguistic auditory signals with novel visual shapes. Importantly, our novel task shares with linguistic processes such as reading acquisition the need to associate sounds with arbitrary shapes. Yet, rather than phonemes or syllables, it uses novel environmental sounds - therefore limiting direct reliance on linguistic abilities. Five-year-old French-speaking children (N = 76, 39 girls) were assessed individually in our novel audio-visual associative task, as well as in a number of other cognitive tasks evaluating linguistic abilities and executive functions. We found phonological awareness and language comprehension to be related to scores in the audio-visual associative task, while no correlation with executive functions was observed. These results underscore a key relation between foundational language competencies and audio-visual associative learning, even in the absence of linguistic input in the associative task.
Collapse
Affiliation(s)
- Irene Altarelli
- Cognitive Neuroimaging Unit U992, INSERM, CEA DRF/Institut Joliot, Université Paris-Sud, Université Paris-Saclay, NeuroSpin Center, Gif/Yvette, France.,Faculty of Psychology and Education Sciences, University of Geneva, Geneva, Switzerland.,CNRS UMR 8240, Laboratory for the Psychology of Child Development and Education (LaPsyDE), University Paris Descartes, Université de Paris, Paris, France
| | - Ghislaine Dehaene-Lambertz
- Cognitive Neuroimaging Unit U992, INSERM, CEA DRF/Institut Joliot, Université Paris-Sud, Université Paris-Saclay, NeuroSpin Center, Gif/Yvette, France
| | - Daphne Bavelier
- Faculty of Psychology and Education Sciences, University of Geneva, Geneva, Switzerland
| |
Collapse
|
24
|
Kemény F, Lukács Á. Sequence in a sequence: Learning of auditory but not visual patterns within a multimodal sequence. Acta Psychol (Amst) 2019; 199:102905. [PMID: 31404742 DOI: 10.1016/j.actpsy.2019.102905] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2019] [Revised: 06/05/2019] [Accepted: 08/04/2019] [Indexed: 11/29/2022] Open
Abstract
The current study investigates whether a unimodal visual and a unimodal auditory sequence is learned separately in a multimodal learning situation. In two experiments participants faced a modified version of the Serial Reaction-Time task, in which auditory and visual elements followed each other, forming a multimodal sequence as well as two unimodal sequences. Learning of both the multimodal and the unimodal sequences were tested. Results showed evidence of multimodal sequence learning. The unimodal sequence formed by the auditory stimuli was also learned, while participants did not acquire the concurrent visual sequence. The experiments argue for a domain- and modality-general sequence learning mechanism that is more sensitive to auditory and response-based information than to visual information.
Collapse
Affiliation(s)
- Ferenc Kemény
- Institute of Psychology, University of Graz, Graz, Austria; Department of Cognitive Science, Budapest University of Technology and Economics, Budapest, Hungary.
| | - Ágnes Lukács
- Department of Cognitive Science, Budapest University of Technology and Economics, Budapest, Hungary; HAS-BME Lendület Language Acquisition Research Group, Budapest, Hungary
| |
Collapse
|
25
|
van Laarhoven T, Stekelenburg JJ, Eussen MLJM, Vroomen J. Electrophysiological alterations in motor-auditory predictive coding in autism spectrum disorder. Autism Res 2019; 12:589-599. [PMID: 30801964 PMCID: PMC6593426 DOI: 10.1002/aur.2087] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2018] [Revised: 12/18/2018] [Accepted: 02/05/2019] [Indexed: 12/28/2022]
Abstract
The amplitude of the auditory N1 component of the event‐related potential (ERP) is typically attenuated for self‐initiated sounds, compared to sounds with identical acoustic and temporal features that are triggered externally. This effect has been ascribed to internal forward models predicting the sensory consequences of one's own motor actions. The predictive coding account of autistic symptomatology states that individuals with autism spectrum disorder (ASD) have difficulties anticipating upcoming sensory stimulation due to a decreased ability to infer the probabilistic structure of their environment. Without precise internal forward prediction models to rely on, perception in ASD could be less affected by prior expectations and more driven by sensory input. Following this reasoning, one would expect diminished attenuation of the auditory N1 due to self‐initiation in individuals with ASD. Here, we tested this hypothesis by comparing the neural response to self‐ versus externally‐initiated tones between a group of individuals with ASD and a group of age matched neurotypical controls. ERPs evoked by tones initiated via button‐presses were compared with ERPs evoked by the same tones replayed at identical pace. Significant N1 attenuation effects were only found in the TD group. Self‐initiation of the tones did not attenuate the auditory N1 in the ASD group, indicating that they may be unable to anticipate the auditory sensory consequences of their own motor actions. These results show that individuals with ASD have alterations in sensory attenuation of self‐initiated sounds, and support the notion of impaired predictive coding as a core deficit underlying autistic symptomatology. Autism Res 2019, 12: 589–599. © 2019 The Authors. Autism Research published by International Society for Autism Research published by Wiley Periodicals, Inc. Lay Summary Many individuals with ASD experience difficulties in processing sensory information (for example, increased sensitivity to sound). Here we show that these difficulties may be related to an inability to anticipate upcoming sensory stimulation. Our findings contribute to a better understanding of the neural mechanisms underlying the different sensory perception experienced by individuals with ASD.
Collapse
Affiliation(s)
- Thijs van Laarhoven
- Department of Cognitive Neuropsychology, Tilburg University, 5000 LE Tilburg, The Netherlands
| | - Jeroen J Stekelenburg
- Department of Cognitive Neuropsychology, Tilburg University, 5000 LE Tilburg, The Netherlands
| | - Mart L J M Eussen
- Department of Child and Adolescent Psychiatry, Yulius Mental Health Organization, Dordrecht, The Netherlands.,Department of Autism, Yulius Mental Health Organization, Dordrecht, The Netherlands
| | - Jean Vroomen
- Department of Cognitive Neuropsychology, Tilburg University, 5000 LE Tilburg, The Netherlands
| |
Collapse
|
26
|
Forest TA, Lichtenfeld A, Alvarez B, Finn AS. Superior learning in synesthetes: Consistent grapheme-color associations facilitate statistical learning. Cognition 2019; 186:72-81. [PMID: 30763803 DOI: 10.1016/j.cognition.2019.02.003] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2017] [Revised: 01/30/2019] [Accepted: 02/01/2019] [Indexed: 12/11/2022]
Abstract
In synesthesia activation in one sensory domain, such as smell or sound, triggers an involuntary and unusual secondary sensory or cognitive experience. In the present study, we ask whether the added sensory experience of synesthesia can aid statistical learning-the ability to track environmental regularities in order to segment continuous information. To investigate this, we measured statistical learning outcomes, using an aurally presented artificial language, in two groups of synesthetes alongside controls and simulated the multimodal experience of synesthesia in non-synesthetes. One group of synesthetes exclusively had grapheme-color (GC) synesthesia, in which the experience of color is automatically triggered by exposure to written or spoken graphemes. The other group had both grapheme-color and sound-color (SC+) synesthesia, in which the experience of color is also triggered by the waveform properties of a voice, such as pitch, timbre, and/or musical chords. Unlike GC-only synesthetes, the experience of color in the SC+ group is not perfectly consistent with the statistics that signal word boundaries. We showed that GC-only synesthetes outperformed both non-synesthetes and SC+ synesthetes, likely because the visual concurrents for GC-only synesthetes are highly consistent with the artificial language. We further observed that our simulations of GC synesthesia, but not SC+ synesthesia produced superior statistical learning, showing that synesthesia likely boosts learning outcomes by providing a consistent secondary cue. Findings are discussed with regard to how multimodal experience can improve learning, with the present data indicating that this boost is more likely to occur through explicit, as opposed to implicit, learning systems.
Collapse
Affiliation(s)
- Tess Allegra Forest
- Department of Psychology, University of Toronto, 100 St. George Street, 4th Floor, Sidney Smith Hall, Toronto, ON M5S 3G3, Canada
| | - Alessandra Lichtenfeld
- Department of Psychology, University of California, Berkeley, Room 3210 Tolman Hall #1650, Berkeley, CA 94720-1650, USA
| | - Bryan Alvarez
- Department of Psychology, University of California, Berkeley, Room 3210 Tolman Hall #1650, Berkeley, CA 94720-1650, USA
| | - Amy S Finn
- Department of Psychology, University of Toronto, 100 St. George Street, 4th Floor, Sidney Smith Hall, Toronto, ON M5S 3G3, Canada.
| |
Collapse
|
27
|
Multimodal Interaction of Contextual and Non-Contextual Sound and Haptics in Virtual Simulations. INFORMATICS-BASEL 2018. [DOI: 10.3390/informatics5040043] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Touch plays a fundamental role in our daily interactions, allowing us to interact with and perceive objects and their spatial properties. Despite its importance in the real-world, touch is often ignored in virtual environments. However, accurately simulating the sense of touch is difficult, requiring the use of high-fidelity haptic devices that are cost-prohibitive. Lower fidelity consumer-level haptic devices are becoming more widespread, yet are generally limited in perceived fidelity and the range of motion (degrees of freedom) required to realistically simulate many tasks. Studies into sound and vision suggest that the presence or absence of sound can influence task performance. Here, we explore whether the presence or absence of contextually relevant sound cues influences the performance of a simple haptic drilling task. Although the results of this study do not show any statistically significant difference in task performance with general (task-irrelevant) sound, we discuss how this is a necessary step in understanding the role of sound on haptic perception.
Collapse
|
28
|
Milne AE, Petkov CI, Wilson B. Auditory and Visual Sequence Learning in Humans and Monkeys using an Artificial Grammar Learning Paradigm. Neuroscience 2018; 389:104-117. [PMID: 28687306 PMCID: PMC6278909 DOI: 10.1016/j.neuroscience.2017.06.059] [Citation(s) in RCA: 22] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2017] [Revised: 06/26/2017] [Accepted: 06/27/2017] [Indexed: 11/30/2022]
Abstract
Language flexibly supports the human ability to communicate using different sensory modalities, such as writing and reading in the visual modality and speaking and listening in the auditory domain. Although it has been argued that nonhuman primate communication abilities are inherently multisensory, direct behavioural comparisons between human and nonhuman primates are scant. Artificial grammar learning (AGL) tasks and statistical learning experiments can be used to emulate ordering relationships between words in a sentence. However, previous comparative work using such paradigms has primarily investigated sequence learning within a single sensory modality. We used an AGL paradigm to evaluate how humans and macaque monkeys learn and respond to identically structured sequences of either auditory or visual stimuli. In the auditory and visual experiments, we found that both species were sensitive to the ordering relationships between elements in the sequences. Moreover, the humans and monkeys produced largely similar response patterns to the visual and auditory sequences, indicating that the sequences are processed in comparable ways across the sensory modalities. These results provide evidence that human sequence processing abilities stem from an evolutionarily conserved capacity that appears to operate comparably across the sensory modalities in both human and nonhuman primates. The findings set the stage for future neurobiological studies to investigate the multisensory nature of these sequencing operations in nonhuman primates and how they compare to related processes in humans.
Collapse
Affiliation(s)
- Alice E Milne
- Institute of Neuroscience, Henry Wellcome Building, Newcastle University, Framlington Place, Newcastle upon Tyne NE2 4HH, United Kingdom; Centre for Behaviour and Evolution, Henry Wellcome Building, Newcastle University, Framlington Place, Newcastle upon Tyne NE2 4HH, United Kingdom
| | - Christopher I Petkov
- Institute of Neuroscience, Henry Wellcome Building, Newcastle University, Framlington Place, Newcastle upon Tyne NE2 4HH, United Kingdom; Centre for Behaviour and Evolution, Henry Wellcome Building, Newcastle University, Framlington Place, Newcastle upon Tyne NE2 4HH, United Kingdom.
| | - Benjamin Wilson
- Institute of Neuroscience, Henry Wellcome Building, Newcastle University, Framlington Place, Newcastle upon Tyne NE2 4HH, United Kingdom; Centre for Behaviour and Evolution, Henry Wellcome Building, Newcastle University, Framlington Place, Newcastle upon Tyne NE2 4HH, United Kingdom
| |
Collapse
|
29
|
Neural Mechanisms of Material Perception: Quest on Shitsukan. Neuroscience 2018; 392:329-347. [PMID: 30213767 DOI: 10.1016/j.neuroscience.2018.09.001] [Citation(s) in RCA: 32] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2018] [Revised: 08/13/2018] [Accepted: 09/03/2018] [Indexed: 01/11/2023]
Abstract
In recent years, a growing body of research has addressed the nature and mechanism of material perception. Material perception entails perceiving and recognizing a material, surface quality or internal state of an object based on sensory stimuli such as visual, tactile, and/or auditory sensations. This process is ongoing in every aspect of daily life. We can, for example, easily distinguish whether an object is made of wood or metal, or whether a surface is rough or smooth. Judging whether the ground is wet or dry or whether a fish is fresh also involves material perception. Information obtained through material perception can be used to govern actions toward objects and to make decisions about whether to approach an object or avoid it. Because the physical processes leading to sensory signals related to material perception is complicated, it has been difficult to manipulate experimental stimuli in a rigorous manner. However, that situation is now changing thanks to advances in technology and knowledge in related fields. In this article, we will review what is currently known about the neural mechanisms responsible for material perception. We will show that cortical areas in the ventral visual pathway are strongly involved in material perception. Our main focus is on vision, but every sensory modality is involved in material perception. Information obtained through different sensory modalities is closely linked in material perception. Such cross-modal processing is another important feature of material perception, and will also be covered in this review.
Collapse
|
30
|
|
31
|
Piazza EA, Denison RN, Silver MA. Recent cross-modal statistical learning influences visual perceptual selection. J Vis 2018; 18:1. [PMID: 29497742 PMCID: PMC5837665 DOI: 10.1167/18.3.1] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Incoming sensory signals are often ambiguous and consistent with multiple perceptual interpretations. Information from one sensory modality can help to resolve ambiguity in another modality, but the mechanisms by which multisensory associations come to influence the contents of conscious perception are unclear. We asked whether and how novel statistical information about the coupling between sounds and images influences the early stages of awareness of visual stimuli. We exposed subjects to consistent, arbitrary pairings of sounds and images and then measured the impact of this recent passive statistical learning on subjects' initial conscious perception of a stimulus by employing binocular rivalry, a phenomenon in which incompatible images presented separately to the two eyes result in a perceptual alternation between the two images. On each trial of the rivalry test, subjects were presented with a pair of rivalrous images (one of which had been consistently paired with a specific sound during exposure while the other had not) and an accompanying sound. We found that, at the onset of binocular rivalry, an image was significantly more likely to be perceived, and was perceived for a longer duration, when it was presented with its paired sound than when presented with other sounds. Our results indicate that recently acquired multisensory information helps resolve sensory ambiguity, and they demonstrate that statistical learning is a fast, flexible mechanism that facilitates this process.
Collapse
Affiliation(s)
- Elise A Piazza
- Princeton Neuroscience Institute, Princeton University, Princeton, NJ, USA.,Vision Science Graduate Group, University of California, Berkeley, Berkeley, CA, USA.,Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA, USA
| | - Rachel N Denison
- Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA, USA.,Department of Psychology and Center for Neural Science, New York University, New York, NY, USA
| | - Michael A Silver
- Vision Science Graduate Group, University of California, Berkeley, Berkeley, CA, USA.,Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA, USA.,School of Optometry, University of California, Berkeley, Berkeley, CA, USA
| |
Collapse
|
32
|
Cross-modal and non-monotonic representations of statistical regularity are encoded in local neural response patterns. Neuroimage 2018; 173:509-517. [PMID: 29477440 DOI: 10.1016/j.neuroimage.2018.02.019] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2017] [Revised: 01/30/2018] [Accepted: 02/12/2018] [Indexed: 11/21/2022] Open
Abstract
Current neurobiological models assign a central role to predictive processes calibrated to environmental statistics. Neuroimaging studies examining the encoding of stimulus uncertainty have relied almost exclusively on manipulations in which stimuli were presented in a single sensory modality, and further assumed that neural responses vary monotonically with uncertainty. This has left a gap in theoretical development with respect to two core issues: (i) are there cross-modal brain systems that encode input uncertainty in way that generalizes across sensory modalities, and (ii) are there brain systems that track input uncertainty in a non-monotonic fashion? We used multivariate pattern analysis to address these two issues using auditory, visual and audiovisual inputs. We found signatures of cross-modal encoding in frontoparietal, orbitofrontal, and association cortices using a searchlight cross-classification analysis where classifiers trained to discriminate levels of uncertainty in one modality were tested in another modality. Additionally, we found widespread systems encoding uncertainty non-monotonically using classifiers trained to discriminate intermediate levels of uncertainty from both the highest and lowest uncertainty levels. These findings comprise the first comprehensive report of cross-modal and non-monotonic neural sensitivity to statistical regularities in the environment, and suggest that conventional paradigms testing for monotonic responses to uncertainty in a single sensory modality may have limited generalizability.
Collapse
|
33
|
Hidaka S, Yaguchi A. An Investigation of the Relationships Between Autistic Traits and Crossmodal Correspondences in Typically Developing Adults. Multisens Res 2018; 31:729-751. [PMID: 31264620 DOI: 10.1163/22134808-20181304] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2017] [Accepted: 04/16/2018] [Indexed: 12/28/2022]
Abstract
Autism spectrum disorder (ASD) includes characteristics such as social and behavioral deficits that are considered common across the general population rather than unique to people with the diagnosis. People with ASD are reported to have sensory irregularities, including crossmodal perception. Crossmodal correspondences are phenomena in which arbitrary crossmodal inputs affect behavioral performance. Crossmodal correspondences are considered to be established through associative learning, but the learning cues are considered to differ across the types of correspondences. In order to investigate whether and how ASD traits affect crossmodal associative learning, this study examined the relationships between the magnitude of crossmodal correspondences and the degree of ASD traits among non-diagnosed adults. We found that, among three types of crossmodal correspondences (brightness-loudness, visual size-pitch, and visual location-pitch pairs), the brightness-loudness pair was related with total ASD traits and a subtrait (social skill). The magnitude of newly learned crossmodal associations (the visual apparent motion direction-pitch pair) also showed a relationship with an ASD subtrait (attention switching). These findings demonstrate that there are unique relationships between crossmodal associations and ASD traits, indicating that each ASD trait is differently involved in sensory associative learning.
Collapse
Affiliation(s)
- Souta Hidaka
- Department of Psychology, Rikkyo University, 1-2-26, Kitano, Niiza-shi, Saitama, 352-8558 Japan
| | - Ayako Yaguchi
- Department of Psychology, Rikkyo University, 1-2-26, Kitano, Niiza-shi, Saitama, 352-8558 Japan
| |
Collapse
|
34
|
Andric M, Davis B, Hasson U. Visual cortex signals a mismatch between regularity of auditory and visual streams. Neuroimage 2017; 157:648-659. [DOI: 10.1016/j.neuroimage.2017.05.028] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2017] [Revised: 04/14/2017] [Accepted: 05/15/2017] [Indexed: 10/19/2022] Open
|
35
|
Versace E, Spierings MJ, Caffini M, Ten Cate C, Vallortigara G. Spontaneous generalization of abstract multimodal patterns in young domestic chicks. Anim Cogn 2017; 20:521-529. [PMID: 28260155 DOI: 10.1007/s10071-017-1079-5] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/04/2016] [Revised: 01/02/2017] [Accepted: 02/27/2017] [Indexed: 10/20/2022]
Abstract
From the early stages of life, learning the regularities associated with specific objects is crucial for making sense of experiences. Through filial imprinting, young precocial birds quickly learn the features of their social partners by mere exposure. It is not clear though to what extent chicks can extract abstract patterns of the visual and acoustic stimuli present in the imprinting object, and how they combine them. To investigate this issue, we exposed chicks (Gallus gallus) to three days of visual and acoustic imprinting, using either patterns with two identical items or patterns with two different items, presented visually, acoustically or in both modalities. Next, chicks were given a choice between the familiar and the unfamiliar pattern, present in either the multimodal, visual or acoustic modality. The responses to the novel stimuli were affected by their imprinting experience, and the effect was stronger for chicks imprinted with multimodal patterns than for the other groups. Interestingly, males and females adopted a different strategy, with males more attracted by unfamiliar patterns and females more attracted by familiar patterns. Our data show that chicks can generalize abstract patterns by mere exposure through filial imprinting and that multimodal stimulation is more effective than unimodal stimulation for pattern learning.
Collapse
Affiliation(s)
- Elisabetta Versace
- Center for Mind/Brain Sciences, University of Trento, Piazza della Manifattura 1, 38068, Rovereto, Italy.
| | - Michelle J Spierings
- Behavioural Biology, Institute of Biology Leiden, Leiden University, Leiden, 2300 RA, The Netherlands.,Leiden Institute for Brain and Cognition, Leiden University, Leiden, 2300 RC, The Netherlands
| | - Matteo Caffini
- Center for Mind/Brain Sciences, University of Trento, Piazza della Manifattura 1, 38068, Rovereto, Italy
| | - Carel Ten Cate
- Behavioural Biology, Institute of Biology Leiden, Leiden University, Leiden, 2300 RA, The Netherlands.,Leiden Institute for Brain and Cognition, Leiden University, Leiden, 2300 RC, The Netherlands
| | - Giorgio Vallortigara
- Center for Mind/Brain Sciences, University of Trento, Piazza della Manifattura 1, 38068, Rovereto, Italy
| |
Collapse
|
36
|
Visual statistical learning of temporal structures at different hierarchical levels. Atten Percept Psychophys 2016; 78:1308-23. [PMID: 27068052 DOI: 10.3758/s13414-016-1104-9] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Visual environments are complex. In order to process the complex information provided by visual environments, the visual system adopts strategies to reduce its complexity. One strategy, called visual statistical learning, or VSL, is to extract the statistical regularities from the environment. Another strategy is to use the hierarchical structure of a scene (e.g., the co-occurrence between local and global information). Through a series of experiments, this study investigated whether the utilization of the statistical regularities and the hierarchical structure could work together to reduce the complexity of a scene. In the familiarization phase, the participants were asked to passively view a stream of hierarchical scenes where the shapes were concurrently presented at the local and global levels. At each of the two levels there were temporal regularities among the three shapes, which always appeared in the same order. In the test phase, the participants judged the familiarity between 2 triplets, whose temporal regularities were either preserved or not. We found that the participants extracted the temporal regularities at each of the local and global levels (Experiment 1). The hierarchical structure influenced the ability to extract the temporal regularities (Experiment 2). Specifically, VSL was either enhanced or impaired depending on whether the hierarchical structure was informative or not. In summary, in order to process a complex scene, the visual system flexibly uses statistical regularities and the hierarchical structure of the scene.
Collapse
|
37
|
Goda N, Yokoi I, Tachibana A, Minamimoto T, Komatsu H. Crossmodal Association of Visual and Haptic Material Properties of Objects in the Monkey Ventral Visual Cortex. Curr Biol 2016; 26:928-34. [DOI: 10.1016/j.cub.2016.02.003] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2015] [Revised: 12/15/2015] [Accepted: 02/01/2016] [Indexed: 02/03/2023]
|
38
|
Walk AM, Conway CM. Cross-Domain Statistical-Sequential Dependencies Are Difficult to Learn. Front Psychol 2016; 7:250. [PMID: 26941696 PMCID: PMC4766371 DOI: 10.3389/fpsyg.2016.00250] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2015] [Accepted: 02/08/2016] [Indexed: 11/13/2022] Open
Abstract
Recent studies have demonstrated participants' ability to learn cross-modal associations during statistical learning tasks. However, these studies are all similar in that the cross-modal associations to be learned occur simultaneously, rather than sequentially. In addition, the majority of these studies focused on learning across sensory modalities but not across perceptual categories. To test both cross-modal and cross-categorical learning of sequential dependencies, we used an artificial grammar learning task consisting of a serial stream of auditory and/or visual stimuli containing both within- and cross-domain dependencies. Experiment 1 examined within-modal and cross-modal learning across two sensory modalities (audition and vision). Experiment 2 investigated within-categorical and cross-categorical learning across two perceptual categories within the same sensory modality (e.g., shape and color; tones and non-words). Our results indicated that individuals demonstrated learning of the within-modal and within-categorical but not the cross-modal or cross-categorical dependencies. These results stand in contrast to the previous demonstrations of cross-modal statistical learning, and highlight the presence of modality constraints that limit the effectiveness of learning in a multimodal environment.
Collapse
Affiliation(s)
- Anne M. Walk
- Neurocognitive Kinesiology Lab, Department of Kinesiology and Community Health, University of Illinois at Urbana-Champaign, UrbanaIL, USA
| | | |
Collapse
|
39
|
Antoniou M, Wong PCM. Varying irrelevant phonetic features hinders learning of the feature being trained. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 139:271-8. [PMID: 26827023 PMCID: PMC4714982 DOI: 10.1121/1.4939736] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/16/2015] [Revised: 12/17/2015] [Accepted: 12/23/2015] [Indexed: 06/05/2023]
Abstract
Learning to distinguish nonnative words that differ in a critical phonetic feature can be difficult. Speech training studies typically employ methods that explicitly direct the learner's attention to the relevant nonnative feature to be learned. However, studies on vision have demonstrated that perceptual learning may occur implicitly, by exposing learners to stimulus features, even if they are irrelevant to the task, and it has recently been suggested that this task-irrelevant perceptual learning framework also applies to speech. In this study, subjects took part in a seven-day training regimen to learn to distinguish one of two nonnative features, namely, voice onset time or lexical tone, using explicit training methods consistent with most speech training studies. Critically, half of the subjects were exposed to stimuli that varied not only in the relevant feature, but in the irrelevant feature as well. The results showed that subjects who were trained with stimuli that varied in the relevant feature and held the irrelevant feature constant achieved the best learning outcomes. Varying both features hindered learning and generalization to new stimuli.
Collapse
Affiliation(s)
- Mark Antoniou
- MARCS Institute, Western Sydney University, Locked Bag 1797, Penrith, New South Wales 2751, Australia
| | - Patrick C M Wong
- Department of Linguistics and Modern Languages and Brain and Mind Institute, The Chinese University of Hong Kong, Shatin, N.T., Hong Kong Special Administrative Region, People's Republic of China
| |
Collapse
|
40
|
Hidaka S, Teramoto W, Sugita Y. Spatiotemporal Processing in Crossmodal Interactions for Perception of the External World: A Review. Front Integr Neurosci 2015; 9:62. [PMID: 26733827 PMCID: PMC4686600 DOI: 10.3389/fnint.2015.00062] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2015] [Accepted: 12/03/2015] [Indexed: 11/13/2022] Open
Abstract
Research regarding crossmodal interactions has garnered much interest in the last few decades. A variety of studies have demonstrated that multisensory information (vision, audition, tactile sensation, and so on) can perceptually interact with each other in the spatial and temporal domains. Findings regarding crossmodal interactions in the spatiotemporal domain (i.e., motion processing) have also been reported, with updates in the last few years. In this review, we summarize past and recent findings on spatiotemporal processing in crossmodal interactions regarding perception of the external world. A traditional view regarding crossmodal interactions holds that vision is superior to audition in spatial processing, but audition is dominant over vision in temporal processing. Similarly, vision is considered to have dominant effects over the other sensory modalities (i.e., visual capture) in spatiotemporal processing. However, recent findings demonstrate that sound could have a driving effect on visual motion perception. Moreover, studies regarding perceptual associative learning reported that, after association is established between a sound sequence without spatial information and visual motion information, the sound sequence could trigger visual motion perception. Other sensory information, such as motor action or smell, has also exhibited similar driving effects on visual motion perception. Additionally, recent brain imaging studies demonstrate that similar activation patterns could be observed in several brain areas, including the motion processing areas, between spatiotemporal information from different sensory modalities. Based on these findings, we suggest that multimodal information could mutually interact in spatiotemporal processing in the percept of the external world and that common perceptual and neural underlying mechanisms would exist for spatiotemporal processing.
Collapse
Affiliation(s)
- Souta Hidaka
- Department of Psychology, Rikkyo University Saitama, Japan
| | - Wataru Teramoto
- Department of Psychology, Kumamoto University Kumamoto, Japan
| | - Yoichi Sugita
- Department of Psychology, Waseda University Tokyo, Japan
| |
Collapse
|
41
|
Kafaligonul H, Oluk C. Audiovisual associations alter the perception of low-level visual motion. Front Integr Neurosci 2015; 9:26. [PMID: 25873869 PMCID: PMC4379893 DOI: 10.3389/fnint.2015.00026] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2014] [Accepted: 03/14/2015] [Indexed: 11/13/2022] Open
Abstract
Motion perception is a pervasive nature of vision and is affected by both immediate pattern of sensory inputs and prior experiences acquired through associations. Recently, several studies reported that an association can be established quickly between directions of visual motion and static sounds of distinct frequencies. After the association is formed, sounds are able to change the perceived direction of visual motion. To determine whether such rapidly acquired audiovisual associations and their subsequent influences on visual motion perception are dependent on the involvement of higher-order attentive tracking mechanisms, we designed psychophysical experiments using regular and reverse-phi random dot motions isolating low-level pre-attentive motion processing. Our results show that an association between the directions of low-level visual motion and static sounds can be formed and this audiovisual association alters the subsequent perception of low-level visual motion. These findings support the view that audiovisual associations are not restricted to high-level attention based motion system and early-level visual motion processing has some potential role.
Collapse
Affiliation(s)
- Hulusi Kafaligonul
- National Magnetic Resonance Research Center (UMRAM), Bilkent University Ankara, Turkey
| | - Can Oluk
- Department of Psychology, Bilkent University Ankara, Turkey
| |
Collapse
|
42
|
Smaller = denser, and the brain knows it: natural statistics of object density shape weight expectations. PLoS One 2015; 10:e0119794. [PMID: 25768977 PMCID: PMC4358826 DOI: 10.1371/journal.pone.0119794] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2014] [Accepted: 01/30/2015] [Indexed: 11/19/2022] Open
Abstract
If one nondescript object's volume is twice that of another, is it necessarily twice as heavy? As larger objects are typically heavier than smaller ones, one might assume humans use such heuristics in preparing to lift novel objects if other informative cues (e.g., material, previous lifts) are unavailable. However, it is also known that humans are sensitive to statistical properties of our environments, and that such sensitivity can bias perception. Here we asked whether statistical regularities in properties of liftable, everyday objects would bias human observers' predictions about objects' weight relationships. We developed state-of-the-art computer vision techniques to precisely measure the volume of everyday objects, and also measured their weight. We discovered that for liftable man-made objects, "twice as large" doesn't mean "twice as heavy": Smaller objects are typically denser, following a power function of volume. Interestingly, this "smaller is denser" relationship does not hold for natural or unliftable objects, suggesting some ideal density range for objects designed to be lifted. We then asked human observers to predict weight relationships between novel objects without lifting them; crucially, these weight predictions quantitatively match typical weight relationships shown by similarly-sized objects in everyday environments. These results indicate that the human brain represents the statistics of everyday objects and that this representation can be quantitatively abstracted and applied to novel objects. Finally, that the brain possesses and can use precise knowledge of the nonlinear association between size and weight carries important implications for implementation of forward models of motor control in artificial systems.
Collapse
|
43
|
The effects of association strength and cross-modal correspondence on the development of multimodal stimuli. Atten Percept Psychophys 2014; 77:560-70. [PMID: 25391886 DOI: 10.3758/s13414-014-0794-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
In addition to temporal and spatial contributions, multimodal binding is also influenced by association strength and the congruency between stimulus elements. A paradigm was established in which an audio-visual stimulus consisting of four attributes (two visual, two auditory) was presented, followed by questions regarding the specific nature of two of those attributes. We wanted to know how association strength and congruency would modulate the basic effect that responding to same-modality information (two visual or two auditory) would be easier than retrieving different-modality information (one visual and one auditory). In Experiment 1, association strengths were compared across three conditions: baseline, intramodal (100 % association within modalities, thereby benefiting same-modality retrieval), and intermodal (100 % association between modalities, thereby benefiting different-modality retrieval). Association strength was shown to damage responses to same-modality information during intermodal conditions. In Experiment 2, association strength was manipulated identically, but was combined with cross-modally corresponding stimuli (further benefiting different-modality retrieval). The locus of the effect was again on responses to same-modality information, damaging responding during intermodal conditions but helping responding during intramodal conditions. The potential contributions of association strength and cross-modal congruency in promoting learning between vision and audition are discussed in relation to a potential default within-modality binding mechanism.
Collapse
|
44
|
Abstract
Our environment is richly structured, with objects producing correlated information within and across sensory modalities. A prominent challenge faced by our perceptual system is to learn such regularities. Here, we examined statistical learning and addressed learners' ability to track transitional probabilities between elements in the auditory and visual modalities. Specifically, we investigated whether cross-modal information affects statistical learning within a single modality. Participants were familiarized with a statistically structured modality (e.g., either audition or vision) accompanied by different types of cues in a second modality (e.g., vision or audition). The results revealed that statistical learning within either modality is affected by cross-modal information, with learning being enhanced or reduced according to the type of cue provided in the second modality.
Collapse
|
45
|
Zilber N, Ciuciu P, Gramfort A, Azizi L, van Wassenhove V. Supramodal processing optimizes visual perceptual learning and plasticity. Neuroimage 2014; 93 Pt 1:32-46. [DOI: 10.1016/j.neuroimage.2014.02.017] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2013] [Revised: 02/05/2014] [Accepted: 02/13/2014] [Indexed: 11/25/2022] Open
|
46
|
Robinson CW, Sloutsky VM. When audition dominates vision: evidence from cross-modal statistical learning. Exp Psychol 2014; 60:113-21. [PMID: 23047918 DOI: 10.1027/1618-3169/a000177] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
Abstract
Presenting information to multiple sensory modalities sometimes facilitates and sometimes interferes with processing of this information. Research examining interference effects shows that auditory input often interferes with processing of visual input in young children (i.e., auditory dominance effect), whereas visual input often interferes with auditory processing in adults (i.e., visual dominance effect). The current study used a cross-modal statistical learning task to examine modality dominance in adults. Participants ably learned auditory and visual statistics when auditory and visual sequences were presented unimodally and when auditory and visual sequences were correlated during training. However, increasing task demands resulted in an important asymmetry: Increased task demands attenuated visual statistical learning, while having no effect on auditory statistical learning. These findings are consistent with auditory dominance effects reported in young children and have important implications for our understanding of how sensory modalities interact while learning the structure of cross-modal information.
Collapse
Affiliation(s)
- Christopher W Robinson
- Department of Psychology and Center for Cognitive Science, The Ohio State University, Columbus, OH 43210, USA.
| | | |
Collapse
|
47
|
Mitchel AD, Christiansen MH, Weiss DJ. Multimodal integration in statistical learning: evidence from the McGurk illusion. Front Psychol 2014; 5:407. [PMID: 24904449 PMCID: PMC4032898 DOI: 10.3389/fpsyg.2014.00407] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2014] [Accepted: 04/18/2014] [Indexed: 11/16/2022] Open
Abstract
Recent advances in the field of statistical learning have established that learners are able to track regularities of multimodal stimuli, yet it is unknown whether the statistical computations are performed on integrated representations or on separate, unimodal representations. In the present study, we investigated the ability of adults to integrate audio and visual input during statistical learning. We presented learners with a speech stream synchronized with a video of a speaker’s face. In the critical condition, the visual (e.g., /gi/) and auditory (e.g., /mi/) signals were occasionally incongruent, which we predicted would produce the McGurk illusion, resulting in the perception of an audiovisual syllable (e.g., /ni/). In this way, we used the McGurk illusion to manipulate the underlying statistical structure of the speech streams, such that perception of these illusory syllables facilitated participants’ ability to segment the speech stream. Our results therefore demonstrate that participants can integrate audio and visual input to perceive the McGurk illusion during statistical learning. We interpret our findings as support for modality-interactive accounts of statistical learning.
Collapse
Affiliation(s)
- Aaron D Mitchel
- Department of Psychology and Program in Neuroscience, Bucknell University Lewisburg, PA, USA
| | - Morten H Christiansen
- Department of Psychology, Cornell University Ithaca, NY, USA ; Department of Language and Communication, University of Southern Denmark Odense, Denmark ; Haskins Laboratories, New Haven CT, USA
| | - Daniel J Weiss
- Department of Psychology and Program in Linguistics, Pennsylvania State University, University Park PA, USA
| |
Collapse
|
48
|
An Overview of Virtual Simulation and Serious Gaming for Surgical Education and Training. STUDIES IN COMPUTATIONAL INTELLIGENCE 2014. [DOI: 10.1007/978-3-642-45432-5_14] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/01/2023]
|
49
|
Mitchel AD, Weiss DJ. Visual speech segmentation: using facial cues to locate word boundaries in continuous speech. LANGUAGE AND COGNITIVE PROCESSES 2014; 29:771-780. [PMID: 25018577 PMCID: PMC4091796 DOI: 10.1080/01690965.2013.791703] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
Speech is typically a multimodal phenomenon, yet few studies have focused on the exclusive contributions of visual cues to language acquisition. To address this gap, we investigated whether visual prosodic information can facilitate speech segmentation. Previous research has demonstrated that language learners can use lexical stress and pitch cues to segment speech and that learners can extract this information from talking faces. Thus, we created an artificial speech stream that contained minimal segmentation cues and paired it with two synchronous facial displays in which visual prosody was either informative or uninformative for identifying word boundaries. Across three familiarisation conditions (audio stream alone, facial streams alone, and paired audiovisual), learning occurred only when the facial displays were informative to word boundaries, suggesting that facial cues can help learners solve the early challenges of language acquisition.
Collapse
Affiliation(s)
- Aaron D. Mitchel
- Department of Psychology, Bucknell University, Lewisburg, PA 17837, USA
| | - Daniel J. Weiss
- Department of Psychology and Program in Linguistics, The Pennsylvania State University, 643 Moore Building, University Park, PA 16802, USA
| |
Collapse
|
50
|
Learning to associate auditory and visual stimuli: behavioral and neural mechanisms. Brain Topogr 2013; 28:479-93. [PMID: 24276220 DOI: 10.1007/s10548-013-0333-7] [Citation(s) in RCA: 31] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2013] [Accepted: 11/11/2013] [Indexed: 12/20/2022]
Abstract
The ability to effectively combine sensory inputs across modalities is vital for acquiring a unified percept of events. For example, watching a hammer hit a nail while simultaneously identifying the sound as originating from the event requires the ability to identify spatio-temporal congruencies and statistical regularities. In this study, we applied a reaction time and hazard function measure known as capacity (e.g., Townsend and AshbyCognitive Theory 200-239, 1978) to quantify the extent to which observers learn paired associations between simple auditory and visual patterns in a model theoretic manner. As expected, results showed that learning was associated with an increase in accuracy, but more significantly, an increase in capacity. The aim of this study was to associate capacity measures of multisensory learning, with neural based measures, namely mean global field power (GFP). We observed a co-variation between an increase in capacity, and a decrease in GFP amplitude as learning occurred. This suggests that capacity constitutes a reliable behavioral index of efficient energy expenditure in the neural domain.
Collapse
|