1
|
Méndez JM, Cooper BG, Goller F. Note similarities affect syntactic stability in zebra finches. J Comp Physiol A Neuroethol Sens Neural Behav Physiol 2024:10.1007/s00359-024-01713-6. [PMID: 39133335 DOI: 10.1007/s00359-024-01713-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2024] [Revised: 07/30/2024] [Accepted: 08/05/2024] [Indexed: 08/13/2024]
Abstract
The acquisition of an acoustic template is a fundamental component of vocal imitation learning, which is used to refine innate vocalizations and develop a species-specific song. In the absence of a model, birds fail to develop species typical songs. In zebra finches (Taeniopygia guttata), tutored birds produce songs with a stereotyped sequence of distinct acoustic elements, or notes, which form the song motif. Songs of untutored individuals feature atypical acoustic and temporal structure. Here we studied songs and associated respiratory patterns of tutored and untutored male zebra finches to investigate whether similar acoustic notes influence the sequence of song elements. A subgroup of animals developed songs with multiple acoustically similar notes that are produced with alike respiratory motor gestures. These birds also showed increased syntactic variability in their adult motif. Sequence variability tended to occur near song elements which showed high similarity in acoustic structure and underlying respiratory motor gestures. The duration and depth of the inspirations preceding the syllables where syntactic variation occurred did not allow prediction of the following sequence of notes, suggesting that the varying duration and air requirement of the following expiratory pulse is not predictively encoded in the motor program. This study provides a novel method for calculation of motor/acoustic similarity, and the results of this study suggest that the note is a fundamental acoustic unit in the organization of the motif and could play a role in the neural code for song syntax.
Collapse
Affiliation(s)
- Jorge M Méndez
- Department of Physics and Astronomy, Minnesota State University-Mankato, Mankato, MN, USA.
| | - Brenton G Cooper
- Department of Psychology, Texas Christian University, Fort Worth, TX, USA
| | - Franz Goller
- Department of Biology, University of Utah, Salt Lake City, UT, USA
- Institute of Zoophysiology, University of Münster, Münster, Germany
| |
Collapse
|
2
|
Heim F, Scharff C, Fisher SE, Riebel K, Ten Cate C. Auditory discrimination learning and acoustic cue weighing in female zebra finches with localized FoxP1 knockdowns. J Neurophysiol 2024; 131:950-963. [PMID: 38629163 DOI: 10.1152/jn.00228.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2023] [Revised: 04/07/2024] [Accepted: 04/11/2024] [Indexed: 05/21/2024] Open
Abstract
Rare disruptions of the transcription factor FOXP1 are implicated in a human neurodevelopmental disorder characterized by autism and/or intellectual disability with prominent problems in speech and language abilities. Avian orthologues of this transcription factor are evolutionarily conserved and highly expressed in specific regions of songbird brains, including areas associated with vocal production learning and auditory perception. Here, we investigated possible contributions of FoxP1 to song discrimination and auditory perception in juvenile and adult female zebra finches. They received lentiviral knockdowns of FoxP1 in one of two brain areas involved in auditory stimulus processing, HVC (proper name) or CMM (caudomedial mesopallium). Ninety-six females, distributed over different experimental and control groups were trained to discriminate between two stimulus songs in an operant Go/Nogo paradigm and subsequently tested with an array of stimuli. This made it possible to assess how well they recognized and categorized altered versions of training stimuli and whether localized FoxP1 knockdowns affected the role of different features during discrimination and categorization of song. Although FoxP1 expression was significantly reduced by the knockdowns, neither discrimination of the stimulus songs nor categorization of songs modified in pitch, sequential order of syllables or by reversed playback were affected. Subsequently, we analyzed the full dataset to assess the impact of the different stimulus manipulations for cue weighing in song discrimination. Our findings show that zebra finches rely on multiple parameters for song discrimination, but with relatively more prominent roles for spectral parameters and syllable sequencing as cues for song discrimination.NEW & NOTEWORTHY In humans, mutations of the transcription factor FoxP1 are implicated in speech and language problems. In songbirds, FoxP1 has been linked to male song learning and female preference strength. We found that FoxP1 knockdowns in female HVC and caudomedial mesopallium (CMM) did not alter song discrimination or categorization based on spectral and temporal information. However, this large dataset allowed to validate different cue weights for spectral over temporal information for song recognition.
Collapse
Affiliation(s)
- Fabian Heim
- Institute of Biology, Leiden University, Leiden, The Netherlands
- Language and Genetics Department, Max Planck Institute for Psycholinguistics,Nijmegen, The Netherlands
- Institute of Biology, Freie Universität Berlin, Berlin, Germany
| | | | - Simon E Fisher
- Language and Genetics Department, Max Planck Institute for Psycholinguistics,Nijmegen, The Netherlands
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| | - Katharina Riebel
- Institute of Biology, Leiden University, Leiden, The Netherlands
| | - Carel Ten Cate
- Institute of Biology, Leiden University, Leiden, The Netherlands
- Leiden Institute for Brain and Cognition, Leiden, The Netherlands
| |
Collapse
|
3
|
Macedo-Lima M, Fernández-Vargas M, Remage-Healey L. Social reinforcement guides operant behaviour and auditory learning in a songbird. Anim Behav 2024; 210:127-137. [PMID: 38505105 PMCID: PMC10947183 DOI: 10.1016/j.anbehav.2024.01.026] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/21/2024]
Abstract
Motivation to seek social interactions is inherent to all social species. For instance, even with risk of disease transmission in a recent pandemic, humans sought out frequent in-person social interactions. In other social animals, socialization can be prioritized even over water or food consumption. Zebra finches, Taeniopygia guttata, are highly gregarious songbirds widely used in behavioural and physiological research. Songbirds, like humans, are vocal learners during development, which rely on intense auditory learning. Aside from supporting song learning, auditory learning further supports individual identification, mate choice and outcome associations in songbirds. To study auditory learning in a laboratory setting, studies often employ operant paradigms with food restriction and reinforcement and require complete social isolation, which can result in stress and other unintended physiological consequences for social species. Thus, in this work, we designed an operant behavioural method leveraging the sociality of zebra finches for goal-directed behaviours. Our approach relies on visual social reinforcement, without depriving the animals of food or social contact. Using this task, we found that visual social reinforcement was a strong motivational drive for operant behaviour. Motivation was sensitive to familiarity towards the stimulus animal and higher when engaging with a familiar versus a novel individual. We further show that this tool can be used to assess auditory discrimination learning using either songs or synthetic pure tones as stimuli. As birds gained experience in the task, they developed a strategy to maximize reward acquisition in spite of receiving more punishment, i.e. liberal response bias. Our operant paradigm provides an alternative to tasks using food reinforcement and could be applied to a variety of highly social species, such as rodents and nonhuman primates.
Collapse
Affiliation(s)
- Matheus Macedo-Lima
- Matheus Macedo-Lima is now at the Department of Biology, University of Maryland, College Park, MD, U.S.A
| | - Marcela Fernández-Vargas
- Marcela Fernández-Vargas is now at the Department of Psychology, Neuroscience Program, Colorado College, Colorado Springs, CO, U.S.A
| | - Luke Remage-Healey
- Corresponding author. (L. Remage-Healey)., @HealeyLab, Neuroscience and Behavior Program, Center for Neuroendocrine Studies, University of Massachusetts Amherst, Amherst MA, U.S.A.
| |
Collapse
|
4
|
Santolin C, Zacharaki K, Toro JM, Sebastian-Galles N. Abstract processing of syllabic structures in early infancy. Cognition 2024; 244:105663. [PMID: 38128322 DOI: 10.1016/j.cognition.2023.105663] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2021] [Revised: 11/02/2023] [Accepted: 11/03/2023] [Indexed: 12/23/2023]
Abstract
Syllables are one of the fundamental building blocks of early language acquisition. From birth onwards, infants preferentially segment, process and represent the speech into syllable-sized units, raising the question of what type of computations infants are able to perform on these perceptual units. Syllables are abstract units structured in a way that allows grouping phonemes into sequences. The goal of this research was to investigate 4-to-5-month-old infants' ability to encode the internal structure of syllables, at a target age when the language system is not yet specialized on the sounds and the phonotactics of native languages. We conducted two experiments in which infants were first familiarized to lists of syllables implementing either CVC (consonant-vowel-consonant) or CCV (consonant-consonant-vowel) structures, then presented with new syllables implementing both structures at test. Experiments differ in the degree of phonological similarity between the materials used at familiarization and test. Results show that infants were able to differentiate syllabic structures at test, even when test syllables were implemented by combinations of phonemes that infants did not hear before. Only infants familiarized with CVC syllables discriminated the structures at test, pointing to a processing advantage for CVC over CCV structures. This research shows that, in addition to preferentially processing the speech into syllable-sized units, during the first months of life, infants are also capable of performing fine-grained computations within such units.
Collapse
Affiliation(s)
- Chiara Santolin
- Center for Brain and Cognition, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Carrer Ramon Trias Fargas 25-27, 08005, Barcelona, Spain.
| | - Konstantina Zacharaki
- Center for Brain and Cognition, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Carrer Ramon Trias Fargas 25-27, 08005, Barcelona, Spain; ESADE Business School, Ramon Llull University, Avenida de Pedralbes, 60-62, 08034, Barcelona, Spain
| | - Juan Manuel Toro
- Center for Brain and Cognition, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Carrer Ramon Trias Fargas 25-27, 08005, Barcelona, Spain; Catalan Institution for Research and Advanced Studies (ICREA), Passeig Lluis Companys, 23, 08010, Barcelona, Spain
| | - Nuria Sebastian-Galles
- Center for Brain and Cognition, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Carrer Ramon Trias Fargas 25-27, 08005, Barcelona, Spain
| |
Collapse
|
5
|
Endress AD. Hebbian learning can explain rhythmic neural entrainment to statistical regularities. Dev Sci 2024:e13487. [PMID: 38372153 DOI: 10.1111/desc.13487] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2023] [Revised: 12/26/2023] [Accepted: 01/29/2024] [Indexed: 02/20/2024]
Abstract
In many domains, learners extract recurring units from continuous sequences. For example, in unknown languages, fluent speech is perceived as a continuous signal. Learners need to extract the underlying words from this continuous signal and then memorize them. One prominent candidate mechanism is statistical learning, whereby learners track how predictive syllables (or other items) are of one another. Syllables within the same word predict each other better than syllables straddling word boundaries. But does statistical learning lead to memories of the underlying words-or just to pairwise associations among syllables? Electrophysiological results provide the strongest evidence for the memory view. Electrophysiological responses can be time-locked to statistical word boundaries (e.g., N400s) and show rhythmic activity with a periodicity of word durations. Here, I reproduce such results with a simple Hebbian network. When exposed to statistically structured syllable sequences (and when the underlying words are not excessively long), the network activation is rhythmic with the periodicity of a word duration and activation maxima on word-final syllables. This is because word-final syllables receive more excitation from earlier syllables with which they are associated than less predictable syllables that occur earlier in words. The network is also sensitive to information whose electrophysiological correlates were used to support the encoding of ordinal positions within words. Hebbian learning can thus explain rhythmic neural activity in statistical learning tasks without any memory representations of words. Learners might thus need to rely on cues beyond statistical associations to learn the words of their native language. RESEARCH HIGHLIGHTS: Statistical learning may be utilized to identify recurring units in continuous sequences (e.g., words in fluent speech) but may not generate explicit memory for words. Exposure to statistically structured sequences leads to rhythmic activity with a period of the duration of the underlying units (e.g., words). I show that a memory-less Hebbian network model can reproduce this rhythmic neural activity as well as putative encodings of ordinal positions observed in earlier research. Direct tests are needed to establish whether statistical learning leads to declarative memories for words.
Collapse
Affiliation(s)
- Ansgar D Endress
- Department of Psychology, City, University of London, London, UK
| |
Collapse
|
6
|
Zebra finches (Taeniopygia guttata) demonstrate cognitive flexibility in using phonology and sequence of syllables in auditory discrimination. Anim Cogn 2023:10.1007/s10071-023-01763-4. [PMID: 36934374 DOI: 10.1007/s10071-023-01763-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2022] [Revised: 02/17/2023] [Accepted: 03/01/2023] [Indexed: 03/20/2023]
Abstract
Zebra finches rely mainly on syllable phonology rather than on syllable sequence when they discriminate between two songs. However, they can also learn to discriminate two strings containing the same set of syllables by their sequence. How learning about the phonological characteristics of syllables and their sequence relate to each other and to the composition of the stimuli is still an open question. We compared whether and how the zebra finches' relative sensitivity for syllable phonology and syllable sequence depends on the differences between syllable strings. Two groups of zebra finches were trained in a Go-Left/Go-Right task to discriminate either between two strings in which each string contained a unique set of song syllables ('Different-syllables group') or two strings in which both strings contained the same set of syllables, but in a different sequential order ('Same-syllables group'). We assessed to what extent the birds in the two experimental groups attend to the spectral characteristics and the sequence of the syllables by measuring the responses to test strings consisting of spectral modifications or sequence changes. Our results showed no difference in the number of trials needed to discriminate strings consisting of either different or identical sets of syllables. Both experimental groups attended to changes in spectral features in a similar way, but the group for which both training strings consisted of the same set of syllables responded more strongly to changes in sequence than the group for which the training strings consisted of different sets of syllables. This outcome suggests the presence of an additional learning process to learn about syllable sequence when learning about syllable phonology is not sufficient to discriminate two strings. Our study thus demonstrates that the relative importance of syllable phonology and sequence depends on how these features vary among stimuli. This indicates cognitive flexibility in the acoustic features that songbirds might use in their song recognition.
Collapse
|
7
|
Cuaya LV, Hernández-Pérez R, Boros M, Deme A, Andics A. Speech naturalness detection and language representation in the dog brain. Neuroimage 2021; 248:118811. [PMID: 34906714 DOI: 10.1016/j.neuroimage.2021.118811] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2021] [Revised: 12/08/2021] [Accepted: 12/10/2021] [Indexed: 12/11/2022] Open
Abstract
Family dogs are exposed to a continuous flow of human speech throughout their lives. However, the extent of their abilities in speech perception is unknown. Here, we used functional magnetic resonance imaging (fMRI) to test speech detection and language representation in the dog brain. Dogs (n = 18) listened to natural speech and scrambled speech in a familiar and an unfamiliar language. Speech scrambling distorts auditory regularities specific to speech and to a given language, but keeps spectral voice cues intact. We hypothesized that if dogs can extract auditory regularities of speech, and of a familiar language, then there will be distinct patterns of brain activity for natural speech vs. scrambled speech, and also for familiar vs. unfamiliar language. Using multivoxel pattern analysis (MVPA) we found that bilateral auditory cortical regions represented natural speech and scrambled speech differently; with a better classifier performance in longer-headed dogs in a right auditory region. This neural capacity for speech detection was not based on preferential processing for speech but rather on sensitivity to sound naturalness. Furthermore, in case of natural speech, distinct activity patterns were found for the two languages in the secondary auditory cortex and in the precruciate gyrus; with a greater difference in responses to the familiar and unfamiliar languages in older dogs, indicating a role for the amount of language exposure. No regions represented differently the scrambled versions of the two languages, suggesting that the activity difference between languages in natural speech reflected sensitivity to language-specific regularities rather than to spectral voice cues. These findings suggest that separate cortical regions support speech naturalness detection and language representation in the dog brain.
Collapse
Affiliation(s)
- Laura V Cuaya
- Department of Ethology, Institute of Biology, Eötvös Loránd University, Budapest, Hungary; MTA-ELTE 'Lendület' Neuroethology of Communication Research Group, Hungarian Academy of Sciences - Eötvös Loránd University, Budapest, Hungary.
| | - Raúl Hernández-Pérez
- Department of Ethology, Institute of Biology, Eötvös Loránd University, Budapest, Hungary; MTA-ELTE 'Lendület' Neuroethology of Communication Research Group, Hungarian Academy of Sciences - Eötvös Loránd University, Budapest, Hungary
| | - Marianna Boros
- Department of Ethology, Institute of Biology, Eötvös Loránd University, Budapest, Hungary; MTA-ELTE 'Lendület' Neuroethology of Communication Research Group, Hungarian Academy of Sciences - Eötvös Loránd University, Budapest, Hungary
| | - Andrea Deme
- Department of Applied Linguistics and Phonetics, Faculty of Humanities, Eötvös Loránd University, Budapest, Hungary; MTA-ELTE 'Lendület' Lingual Articulation Research Group, Budapest, Hungary
| | - Attila Andics
- Department of Ethology, Institute of Biology, Eötvös Loránd University, Budapest, Hungary; MTA-ELTE 'Lendület' Neuroethology of Communication Research Group, Hungarian Academy of Sciences - Eötvös Loránd University, Budapest, Hungary
| |
Collapse
|
8
|
Boros M, Magyari L, Török D, Bozsik A, Deme A, Andics A. Neural processes underlying statistical learning for speech segmentation in dogs. Curr Biol 2021; 31:5512-5521.e5. [PMID: 34717832 DOI: 10.1016/j.cub.2021.10.017] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2021] [Revised: 07/23/2021] [Accepted: 10/07/2021] [Indexed: 10/20/2022]
Abstract
To learn words, humans extract statistical regularities from speech. Multiple species use statistical learning also to process speech, but the neural underpinnings of speech segmentation in non-humans remain largely unknown. Here, we investigated computational and neural markers of speech segmentation in dogs, a phylogenetically distant mammal that efficiently navigates humans' social and linguistic environment. Using electroencephalography (EEG), we compared event-related responses (ERPs) for artificial words previously presented in a continuous speech stream with different distributional statistics. Results revealed an early effect (220-470 ms) of transitional probability and a late component (590-790 ms) modulated by both word frequency and transitional probability. Using fMRI, we searched for brain regions sensitive to statistical regularities in speech. Structured speech elicited lower activity in the basal ganglia, a region involved in sequence learning, and repetition enhancement in the auditory cortex. Speech segmentation in dogs, similar to that of humans, involves complex computations, engaging both domain-general and modality-specific brain areas. VIDEO ABSTRACT.
Collapse
Affiliation(s)
- Marianna Boros
- MTA-ELTE "Lendület" Neuroethology of Communication Research Group, Hungarian Academy of Sciences - Eötvös Loránd University, 1117 Budapest, Pázmány Péter sétány 1/C, Hungary; Department of Ethology, Eötvös Loránd University, 1117 Budapest, Pázmány Péter sétány 1/C, Hungary.
| | - Lilla Magyari
- MTA-ELTE "Lendület" Neuroethology of Communication Research Group, Hungarian Academy of Sciences - Eötvös Loránd University, 1117 Budapest, Pázmány Péter sétány 1/C, Hungary; Department of Ethology, Eötvös Loránd University, 1117 Budapest, Pázmány Péter sétány 1/C, Hungary; Norwegian Reading Centre for Reading Education and Research, Faculty of Arts and Education, University of Stavanger, Professor Olav Hanssens vei 10, 4036 Stavanger, Norway
| | - Dávid Török
- MTA-ELTE "Lendület" Neuroethology of Communication Research Group, Hungarian Academy of Sciences - Eötvös Loránd University, 1117 Budapest, Pázmány Péter sétány 1/C, Hungary; Department of Ethology, Eötvös Loránd University, 1117 Budapest, Pázmány Péter sétány 1/C, Hungary
| | - Anett Bozsik
- MTA-ELTE "Lendület" Neuroethology of Communication Research Group, Hungarian Academy of Sciences - Eötvös Loránd University, 1117 Budapest, Pázmány Péter sétány 1/C, Hungary; Department of Anatomy and Histology, University of Veterinary Medicine, 1078 Budapest, István utca 2, Hungary
| | - Andrea Deme
- Department of Applied Linguistics and Phonetics, Eötvös Loránd University, 1088 Budapest, Múzeum krt. 4/A, Hungary; MTA-ELTE "Lendület" Lingual Articulation Research Group, 1088 Budapest, Múzeum krt. 4/A, Hungary
| | - Attila Andics
- MTA-ELTE "Lendület" Neuroethology of Communication Research Group, Hungarian Academy of Sciences - Eötvös Loránd University, 1117 Budapest, Pázmány Péter sétány 1/C, Hungary; Department of Ethology, Eötvös Loránd University, 1117 Budapest, Pázmány Péter sétány 1/C, Hungary.
| |
Collapse
|
9
|
Endress AD, Johnson SP. When forgetting fosters learning: A neural network model for statistical learning. Cognition 2021; 213:104621. [PMID: 33608130 DOI: 10.1016/j.cognition.2021.104621] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2020] [Revised: 12/19/2020] [Accepted: 01/28/2021] [Indexed: 11/28/2022]
Abstract
Learning often requires splitting continuous signals into recurring units, such as the discrete words constituting fluent speech; these units then need to be encoded in memory. A prominent candidate mechanism involves statistical learning of co-occurrence statistics like transitional probabilities (TPs), reflecting the idea that items from the same unit (e.g., syllables within a word) predict each other better than items from different units. TP computations are surprisingly flexible and sophisticated. Humans are sensitive to forward and backward TPs, compute TPs between adjacent items and longer-distance items, and even recognize TPs in novel units. We explain these hallmarks of statistical learning with a simple model with tunable, Hebbian excitatory connections and inhibitory interactions controlling the overall activation. With weak forgetting, activations are long-lasting, yielding associations among all items; with strong forgetting, no associations ensue as activations do not outlast stimuli; with intermediate forgetting, the network reproduces the hallmarks above. Forgetting thus is a key determinant of these sophisticated learning abilities. Further, in line with earlier dissociations between statistical learning and memory encoding, our model reproduces the hallmarks of statistical learning in the absence of a memory store in which items could be placed.
Collapse
|
10
|
Different mechanisms underlie implicit visual statistical learning in honey bees and humans. Proc Natl Acad Sci U S A 2020; 117:25923-25934. [PMID: 32989162 DOI: 10.1073/pnas.1919387117] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
The ability of developing complex internal representations of the environment is considered a crucial antecedent to the emergence of humans' higher cognitive functions. Yet it is an open question whether there is any fundamental difference in how humans and other good visual learner species naturally encode aspects of novel visual scenes. Using the same modified visual statistical learning paradigm and multielement stimuli, we investigated how human adults and honey bees (Apis mellifera) encode spontaneously, without dedicated training, various statistical properties of novel visual scenes. We found that, similarly to humans, honey bees automatically develop a complex internal representation of their visual environment that evolves with accumulation of new evidence even without a targeted reinforcement. In particular, with more experience, they shift from being sensitive to statistics of only elemental features of the scenes to relying on co-occurrence frequencies of elements while losing their sensitivity to elemental frequencies, but they never encode automatically the predictivity of elements. In contrast, humans involuntarily develop an internal representation that includes single-element and co-occurrence statistics, as well as information about the predictivity between elements. Importantly, capturing human visual learning results requires a probabilistic chunk-learning model, whereas a simple fragment-based memory-trace model that counts occurrence summary statistics is sufficient to replicate honey bees' learning behavior. Thus, humans' sophisticated encoding of sensory stimuli that provides intrinsic sensitivity to predictive information might be one of the fundamental prerequisites of developing higher cognitive abilities.
Collapse
|
11
|
Endress AD, Slone LK, Johnson SP. Statistical learning and memory. Cognition 2020; 204:104346. [PMID: 32615468 DOI: 10.1016/j.cognition.2020.104346] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2019] [Revised: 05/12/2020] [Accepted: 05/30/2020] [Indexed: 01/01/2023]
Abstract
Learners often need to identify and remember recurring units in continuous sequences, but the underlying mechanisms are debated. A particularly prominent candidate mechanism relies on distributional statistics such as Transitional Probabilities (TPs). However, it is unclear what the outputs of statistical segmentation mechanisms are, and if learners store these outputs as discrete chunks in memory. We critically review the evidence for the possibility that statistically coherent items are stored in memory and outline difficulties in interpreting past research. We use Slone and Johnson's (2018) experiments as a case study to show that it is difficult to delineate the different mechanisms learners might use to solve a learning problem. Slone and Johnson (2018) reported that 8-month-old infants learned coherent chunks of shapes in visual sequences. Here, we describe an alternate interpretation of their findings based on a multiple-cue integration perspective. First, when multiple cues to statistical structure were available, infants' looking behavior seemed to track with the strength of the strongest one - backward TPs, suggesting that infants process multiple cues simultaneously and select the strongest one. Second, like adults, infants are exquisitely sensitive to chunks, but may require multiple cues to extract them. In Slone and Johnson's (2018) experiments, these cues were provided by immediate chunk repetitions during familiarization. Accordingly, infants showed strongest evidence of chunking following familiarization sequences in which immediate repetitions were more frequent. These interpretations provide a strong argument for infants' processing of multiple cues and the potential importance of multiple cues for chunk recognition in infancy.
Collapse
Affiliation(s)
- Ansgar D Endress
- Department of Psychology, City, University of London, United Kingdom.
| | - Lauren K Slone
- Department of Psychological and Brain Sciences, Indiana University, Bloomington, United States; Department of Psychology, Hope College, Holland, United States
| | - Scott P Johnson
- Department of Psychology, University of California, Los Angeles, United States
| |
Collapse
|
12
|
Wilson B, Spierings M, Ravignani A, Mueller JL, Mintz TH, Wijnen F, van der Kant A, Smith K, Rey A. Non-adjacent Dependency Learning in Humans and Other Animals. Top Cogn Sci 2020; 12:843-858. [PMID: 32729673 PMCID: PMC7496455 DOI: 10.1111/tops.12381] [Citation(s) in RCA: 37] [Impact Index Per Article: 9.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2018] [Revised: 05/22/2018] [Accepted: 05/30/2018] [Indexed: 11/28/2022]
Abstract
Learning and processing natural language requires the ability to track syntactic relationships between words and phrases in a sentence, which are often separated by intervening material. These nonadjacent dependencies can be studied using artificial grammar learning paradigms and structured sequence processing tasks. These approaches have been used to demonstrate that human adults, infants and some nonhuman animals are able to detect and learn dependencies between nonadjacent elements within a sequence. However, learning nonadjacent dependencies appears to be more cognitively demanding than detecting dependencies between adjacent elements, and only occurs in certain circumstances. In this review, we discuss different types of nonadjacent dependencies in language and in artificial grammar learning experiments, and how these differences might impact learning. We summarize different types of perceptual cues that facilitate learning, by highlighting the relationship between dependent elements bringing them closer together either physically, attentionally, or perceptually. Finally, we review artificial grammar learning experiments in human adults, infants, and nonhuman animals, and discuss how similarities and differences observed across these groups can provide insights into how language is learned across development and how these language-related abilities might have evolved.
Collapse
Affiliation(s)
| | | | - Andrea Ravignani
- Research DepartmentSealcentre Pieterburen
- Artificial Intelligence LabVrije Universiteit Brussel
| | | | - Toben H. Mintz
- Departments of Psychology and LinguisticsUniversity of Southern California
| | - Frank Wijnen
- Utrecht Institute of Linguistics OTSUtrecht University
| | | | - Kenny Smith
- Centre for Language EvolutionUniversity of Edinburgh
| | | |
Collapse
|
13
|
Trotter AS, Monaghan P, Beckers GJL, Christiansen MH. Exploring Variation Between Artificial Grammar Learning Experiments: Outlining a Meta-Analysis Approach. Top Cogn Sci 2020; 12:875-893. [PMID: 31495072 PMCID: PMC7496870 DOI: 10.1111/tops.12454] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2018] [Revised: 07/14/2019] [Accepted: 07/25/2019] [Indexed: 11/30/2022]
Abstract
Artificial grammar learning (AGL) has become an important tool used to understand aspects of human language learning and whether the abilities underlying learning may be unique to humans or found in other species. Successful learning is typically assumed when human or animal participants are able to distinguish stimuli generated by the grammar from those that are not at a level better than chance. However, the question remains as to what subjects actually learn in these experiments. Previous studies of AGL have frequently introduced multiple potential contributors to performance in the training and testing stimuli, but meta-analysis techniques now enable us to consider these multiple information sources for their contribution to learning-enabling intended and unintended structures to be assessed simultaneously. We present a blueprint for meta-analysis approaches to appraise the effect of learning in human and other animal studies for a series of artificial grammar learning experiments, focusing on studies that examine auditory and visual modalities. We identify a series of variables that differ across these studies, focusing on both structural and surface properties of the grammar, and characteristics of training and test regimes, and provide a first step in assessing the relative contribution of these design features of artificial grammars as well as species-specific effects for learning.
Collapse
Affiliation(s)
- Antony S. Trotter
- Department of Speech, Hearing & Phonetic SciencesUniversity College London
| | - Padraic Monaghan
- Department of PsychologyLancaster University
- Department of EnglishUniversity of Amsterdam
| | - Gabriël J. L. Beckers
- Department of Psychology, Cognitive Neurobiology and Helmholtz InstituteUtrecht University
| | - Morten H. Christiansen
- Department of PsychologyCornell University
- Interacting Minds Centre and School of Communication and CultureAarhus University
- Haskins Laboratories
| |
Collapse
|
14
|
Petkov CI, ten Cate C. Structured Sequence Learning: Animal Abilities, Cognitive Operations, and Language Evolution. Top Cogn Sci 2020; 12:828-842. [PMID: 31359600 PMCID: PMC7537567 DOI: 10.1111/tops.12444] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2018] [Revised: 06/20/2019] [Accepted: 06/20/2019] [Indexed: 11/28/2022]
Abstract
Human language is a salient example of a neurocognitive system that is specialized to process complex dependencies between sensory events distributed in time, yet how this system evolved and specialized remains unclear. Artificial Grammar Learning (AGL) studies have generated a wealth of insights into how human adults and infants process different types of sequencing dependencies of varying complexity. The AGL paradigm has also been adopted to examine the sequence processing abilities of nonhuman animals. We critically evaluate this growing literature in species ranging from mammals (primates and rats) to birds (pigeons, songbirds, and parrots) considering also cross-species comparisons. The findings are contrasted with seminal studies in human infants that motivated the work in nonhuman animals. This synopsis identifies advances in knowledge and where uncertainty remains regarding the various strategies that nonhuman animals can adopt for processing sequencing dependencies. The paucity of evidence in the few species studied to date and the need for follow-up experiments indicate that we do not yet understand the limits of animal sequence processing capacities and thereby the evolutionary pattern. This vibrant, yet still budding, field of research carries substantial promise for advancing knowledge on animal abilities, cognitive substrates, and language evolution.
Collapse
|
15
|
Mizuhara T, Okanoya K. Do songbirds hear songs syllable by syllable? Behav Processes 2020; 174:104089. [PMID: 32105758 DOI: 10.1016/j.beproc.2020.104089] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2019] [Revised: 01/30/2020] [Accepted: 02/23/2020] [Indexed: 12/29/2022]
Abstract
Songbirds as vocal learners have been one of the most popular model species to investigate the biological prerequisite to human language. Their songs consist of syllables, which appear as pulse trains in sound spectrograms. When describing the song sequence, researchers consider the syllable to be the unit of the song. Moreover, artificial grammar learning studies asking whether songbirds recognize structural regularities observed in human language often design stimuli using song syllables as components. However, whether syllables are perceptual units is yet to be determined. We found that Bengalese finches, a species of songbird, responded significantly less to one specific syllable when it was temporally placed close to the preceding syllable. The proximity, or silent interval was within the range of what is produced in the natural songs of both Bengalese and zebra finches, and what has been used in other artificial grammar learning studies using zebra finches. Our results suggest the need for a reinterpretation of the description of birdsong structure and of previous artificial grammar learning studies.
Collapse
Affiliation(s)
- Tomoko Mizuhara
- Department of Life Sciences, Graduate School of Arts and Sciences, The University of Tokyo, Meguro, Tokyo 153-8902, Japan.
| | - Kazuo Okanoya
- Department of Life Sciences, Graduate School of Arts and Sciences, The University of Tokyo, Meguro, Tokyo 153-8902, Japan.
| |
Collapse
|
16
|
Statistical learning for vocal sequence acquisition in a songbird. Sci Rep 2020; 10:2248. [PMID: 32041978 PMCID: PMC7010765 DOI: 10.1038/s41598-020-58983-8] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2018] [Accepted: 01/17/2020] [Indexed: 01/31/2023] Open
Abstract
Birdsong is a learned communicative behavior that consists of discrete acoustic elements (“syllables”) that are sequenced in a controlled manner. While the learning of the acoustic structure of syllables has been extensively studied, relatively little is known about sequence learning in songbirds. Statistical learning could contribute to the acquisition of vocal sequences, and we investigated the nature and extent of sequence learning at various levels of song organization in the Bengalese finch, Lonchura striata var. domestica. We found that, under semi-natural conditions, pupils (sons) significantly reproduced the sequence statistics of their tutor’s (father’s) songs at multiple levels of organization (e.g., syllable repertoire, prevalence, and transitions). For example, the probability of syllable transitions at “branch points” (relatively complex sequences that are followed by multiple types of transitions) were significantly correlated between the songs of tutors and pupils. We confirmed the contribution of learning to sequence similarities between fathers and sons by experimentally tutoring juvenile Bengalese finches with the songs of unrelated tutors. We also discovered that the extent and fidelity of sequence similarities between tutors and pupils were significantly predicted by the prevalence of sequences in the tutor’s song and that distinct types of sequence modifications (e.g., syllable additions or deletions) followed distinct patterns. Taken together, these data provide compelling support for the role of statistical learning in vocal production learning and identify factors that could modulate the extent of vocal sequence learning.
Collapse
|
17
|
Fishbein AR, Idsardi WJ, Ball GF, Dooling RJ. Sound sequences in birdsong: how much do birds really care? Philos Trans R Soc Lond B Biol Sci 2019; 375:20190044. [PMID: 31735149 DOI: 10.1098/rstb.2019.0044] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
The complex and melodic nature of many birds' songs has raised interest in potential parallels between avian vocal sequences and human speech. The similarities between birdsong and speech in production and learning are well established, but surprisingly little is known about how birds perceive song sequences. One popular laboratory songbird, the zebra finch (Taeniopygia guttata), has recently attracted attention as an avian model for human speech, in part because the male learns to produce the individual elements in its song motif in a fixed sequence. But psychoacoustic evidence shows that adult zebra finches are relatively insensitive to the sequential features of song syllables. Instead, zebra finches and other birds seem to be exquisitely sensitive to the acoustic details of individual syllables to a degree that is beyond human hearing capacity. Based on these findings, we present a finite-state model of zebra finch perception of song syllable sequences and discuss the rich informational capacity of their vocal system. Furthermore, we highlight the abilities of budgerigars (Melopsittacus undulatus), a parrot species, to hear sequential features better than zebra finches and suggest that neurophysiological investigations comparing these species could prove fruitful for uncovering neural mechanisms for auditory sequence perception in human speech. This article is part of the theme issue 'What can animal communication teach us about human language?'
Collapse
Affiliation(s)
- Adam R Fishbein
- Psychology Department, University of Maryland, 4094 Campus Drive, College Park, MD 20742, USA.,Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD 20742, USA
| | - William J Idsardi
- Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD 20742, USA.,Linguistics Department, University of Maryland, 1401 Marie Mount Hall, College Park, MD 20742, USA
| | - Gregory F Ball
- Psychology Department, University of Maryland, 4094 Campus Drive, College Park, MD 20742, USA.,Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD 20742, USA
| | - Robert J Dooling
- Psychology Department, University of Maryland, 4094 Campus Drive, College Park, MD 20742, USA.,Neuroscience and Cognitive Science Program, University of Maryland, College Park, MD 20742, USA
| |
Collapse
|
18
|
Neuronal Encoding in a High-Level Auditory Area: From Sequential Order of Elements to Grammatical Structure. J Neurosci 2019; 39:6150-6161. [PMID: 31147525 DOI: 10.1523/jneurosci.2767-18.2019] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2018] [Revised: 04/24/2019] [Accepted: 04/24/2019] [Indexed: 12/27/2022] Open
Abstract
Sensitivity to the sequential structure of communication sounds is fundamental not only for language comprehension in humans but also for song recognition in songbirds. By quantifying single-unit responses, we first assessed whether the sequential order of song elements, called syllables, in conspecific songs is encoded in a secondary auditory cortex-like region of the zebra finch brain. Based on a habituation/dishabituation paradigm, we show that, after multiple repetitions of the same conspecific song, rearranging syllable order reinstated strong responses. A large proportion of neurons showed sensitivity to song context in which syllables occurred providing support for the nonlinear processing of syllable sequences. Sensitivity to the temporal order of items within a sequence should enable learning its underlying structure, an ability considered a core mechanism of the human language faculty. We show that repetitions of songs that were ordered according to a specific grammatical structure (i.e., ABAB or AABB structures; A and B denoting song syllables) led to different responses in both anesthetized and awake birds. Once responses were decreased due to song repetitions, the transition from one structure to the other could affect the firing rates and/or the spike patterns. Our results suggest that detection was based on local differences rather than encoding of the global song structure as a whole. Our study demonstrates that a high-level auditory region provides neuronal mechanisms to help discriminate stimuli that differ in their sequential structure.SIGNIFICANCE STATEMENT Sequence processing has been proposed as a potential precursor of language syntax. As a sequencing operation, the encoding of the temporal order of items within a sequence may help in recognition of relationships between adjacent items and in learning the underlying structure. Taking advantage of the stimulus-specific adaptation phenomenon observed in a high-level auditory region of the zebra finch brain, we addressed this question at the neuronal level. Reordering elements within conspecific songs reinstated robust responses. Neurons also detected changes in the structure of artificial songs, and this detection depended on local transitions between adjacent or nonadjacent syllables. These findings establish the songbird as a model system for deciphering the mechanisms underlying sequence processing at the single-cell level.
Collapse
|
19
|
Maheu M, Dehaene S, Meyniel F. Brain signatures of a multiscale process of sequence learning in humans. eLife 2019; 8:41541. [PMID: 30714904 PMCID: PMC6361584 DOI: 10.7554/elife.41541] [Citation(s) in RCA: 46] [Impact Index Per Article: 9.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2018] [Accepted: 01/18/2019] [Indexed: 01/08/2023] Open
Abstract
Extracting the temporal structure of sequences of events is crucial for perception, decision-making, and language processing. Here, we investigate the mechanisms by which the brain acquires knowledge of sequences and the possibility that successive brain responses reflect the progressive extraction of sequence statistics at different timescales. We measured brain activity using magnetoencephalography in humans exposed to auditory sequences with various statistical regularities, and we modeled this activity as theoretical surprise levels using several learning models. Successive brain waves related to different types of statistical inferences. Early post-stimulus brain waves denoted a sensitivity to a simple statistic, the frequency of items estimated over a long timescale (habituation). Mid-latency and late brain waves conformed qualitatively and quantitatively to the computational properties of a more complex inference: the learning of recent transition probabilities. Our findings thus support the existence of multiple computational systems for sequence processing involving statistical inferences at multiple scales.
Collapse
Affiliation(s)
- Maxime Maheu
- Cognitive Neuroimaging Unit, CEA DRF/JOLIOT, INSERM, Université Paris-Sud, Université Paris-Saclay, NeuroSpin center, Gif-sur-Yvette, France.,Université Paris Descartes, Sorbonne Paris Cité, Paris, France
| | - Stanislas Dehaene
- Cognitive Neuroimaging Unit, CEA DRF/JOLIOT, INSERM, Université Paris-Sud, Université Paris-Saclay, NeuroSpin center, Gif-sur-Yvette, France.,Collège de France, Paris, France
| | - Florent Meyniel
- Cognitive Neuroimaging Unit, CEA DRF/JOLIOT, INSERM, Université Paris-Sud, Université Paris-Saclay, NeuroSpin center, Gif-sur-Yvette, France
| |
Collapse
|
20
|
Spontaneous Learning of Visual Structures in Domestic Chicks. Animals (Basel) 2018; 8:ani8080135. [PMID: 30082590 PMCID: PMC6115858 DOI: 10.3390/ani8080135] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2018] [Revised: 07/31/2018] [Accepted: 08/02/2018] [Indexed: 11/17/2022] Open
Abstract
Simple Summary Our aim is to investigate the recognition of the structure of multi-element configurations; one mechanism that supports communicative functions in different species. Cognitive mechanisms involved in this ability might not have evolved specifically for communicative use, but derive from other functions. Thus, it is crucial to study these abilities in species that are not vocal learners and with stimuli from other modalities. We know already that domestic chicks can learn the temporal statistical structure of sequences of visual shapes, however their abilities to encode the spatial structure of visual patterns (configurations composed of multiple visual elements presented simultaneously side-by-side) is much less known. Using filial imprinting learning, we showed that chicks spontaneously recognize the structure of their imprinting stimulus, preferring it to one composed of the same elements in different configurations. Moreover, we found that in their affiliative responses chicks give priority to information located at the stimulus edges, a phenomenon that was so far observed only with temporal sequences. This first evidence of a spontaneous edge bias with spatial stimuli further stresses the importance of studying similarities and differences between the processing of linguistic and nonlinguistic stimuli and of stimuli presented in various sensory modalities. Abstract Effective communication crucially depends on the ability to produce and recognize structured signals, as apparent in language and birdsong. Although it is not clear to what extent similar syntactic-like abilities can be identified in other animals, recently we reported that domestic chicks can learn abstract visual patterns and the statistical structure defined by a temporal sequence of visual shapes. However, little is known about chicks’ ability to process spatial/positional information from visual configurations. Here, we used filial imprinting as an unsupervised learning mechanism to study spontaneous encoding of the structure of a configuration of different shapes. After being exposed to a triplet of shapes (ABC or CAB), chicks could discriminate those triplets from a permutation of the same shapes in different order (CAB or ABC), revealing a sensitivity to the spatial arrangement of the elements. When tested with a fragment taken from the imprinting triplet that followed the familiar adjacency-relationships (AB or BC) vs. one in which the shapes maintained their position with respect to the stimulus edges (AC), chicks revealed a preference for the configuration with familiar edge elements, showing an edge bias previously found only with temporal sequences.
Collapse
|
21
|
Knowles JM, Doupe AJ, Brainard MS. Zebra finches are sensitive to combinations of temporally distributed features in a model of word recognition. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:872. [PMID: 30180710 PMCID: PMC6103769 DOI: 10.1121/1.5050910] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/01/2018] [Accepted: 07/21/2018] [Indexed: 06/08/2023]
Abstract
Discrimination between spoken words composed of overlapping elements, such as "captain" and "captive," relies on sensitivity to unique combinations of prefix and suffix elements that span a "uniqueness point" where the word candidates diverge. To model such combinatorial processing, adult female zebra finches were trained to discriminate between target and distractor syllable sequences that shared overlapping "contextual" prefixes and differed only in their "informative" suffixes. The transition from contextual to informative syllables thus created a uniqueness point analogous to that present between overlapping word candidates, where targets and distractors diverged. It was found that target recognition depended not only on informative syllables, but also on contextual syllables that were shared with distractors. Moreover, the influence of each syllable depended on proximity to the uniqueness point. Birds were then trained birds with targets and distractors that shared both prefix and suffix sequences and could only be discriminated by recognizing unique combinations of those sequences. Birds learned to robustly discriminate target and distractor combinations and maintained significant discrimination when the local transitions from prefix to suffix were disrupted. These findings indicate that birds, like humans, combine information across temporally distributed features, spanning contextual and informative elements, in recognizing and discriminating word-like stimuli.
Collapse
Affiliation(s)
- Jeffrey M Knowles
- Center for Integrative Neuroscience, University of California, San Francisco, 675 Nelson Rising Lane, San Francisco, California 94158, USA
| | - Allison J Doupe
- Center for Integrative Neuroscience, University of California, San Francisco, 675 Nelson Rising Lane, San Francisco, California 94158, USA
| | - Michael S Brainard
- Howard Hughes Medical Institute, University of California, San Francisco, San Francisco, California 94158, USA
| |
Collapse
|
22
|
|
23
|
Santolin C, Saffran JR. Constraints on Statistical Learning Across Species. Trends Cogn Sci 2018; 22:52-63. [PMID: 29150414 PMCID: PMC5777226 DOI: 10.1016/j.tics.2017.10.003] [Citation(s) in RCA: 65] [Impact Index Per Article: 10.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2017] [Revised: 10/13/2017] [Accepted: 10/16/2017] [Indexed: 10/18/2022]
Abstract
Both human and nonhuman organisms are sensitive to statistical regularities in sensory inputs that support functions including communication, visual processing, and sequence learning. One of the issues faced by comparative research in this field is the lack of a comprehensive theory to explain the relevance of statistical learning across distinct ecological niches. In the current review we interpret cross-species research on statistical learning based on the perceptual and cognitive mechanisms that characterize the human and nonhuman models under investigation. Considering statistical learning as an essential part of the cognitive architecture of an animal will help to uncover the potential ecological functions of this powerful learning process.
Collapse
Affiliation(s)
- Chiara Santolin
- Center for Brain and Cognition, Universitat Pompeu Fabra, Carrer Ramon Trias Fargas, 25-27, 08005 Barcelona, Spain.
| | - Jenny R Saffran
- Waisman Center, University of Wisconsin-Madison, 1500 Highland Avenue, Madison, WI 53705, USA
| |
Collapse
|
24
|
Abstract
Questions related to the uniqueness of language can only be addressed properly by referring to sound knowledge of the relevant cognitive abilities of nonhuman animals. A key question concerns the nature and extent of animal rule-learning abilities. I discuss two approaches used to assess these abilities. One is comparing the structures of animal vocalizations to linguistic ones, and another is addressing the grammatical rule- and pattern-learning abilities of animals through experiments using artificial grammars. Neither of these approaches has so far provided unambiguous evidence of advanced animal abilities. However, when we consider how animal vocalizations are analyzed, the types of stimuli and tasks that are used in artificial grammar learning experiments, the limited number of species examined, and the groups to which these belong, I argue that the currently available evidence is insufficient to arrive at firm conclusions concerning the limitations of animal grammatical abilities. As a consequence, the gap between human linguistic rule-learning abilities and those of nonhuman animals may be smaller and less clear than is currently assumed. This means that it is still an open question whether a difference in the rule-learning and rule abstraction abilities between animals and humans played the key role in the evolution of language.
Collapse
|
25
|
Abstract
Learning to read involves the acquisition of letter-sound relationships (i.e., decoding skills) and the ability to visually recognize words (i.e., orthographic knowledge). Although decoding skills are clearly human-unique, given they are seated in language, recent research and theory suggest that orthographic processing may derive from the exaptation or recycling of visual circuits that evolved to recognize everyday objects and shapes in our natural environment. An open question is whether orthographic processing is limited to visual circuits that are similar to our own or a product of plasticity common to many vertebrate visual systems. Here we show that pigeons, organisms that separated from humans more than 300 million y ago, process words orthographically. Specifically, we demonstrate that pigeons trained to discriminate words from nonwords picked up on the orthographic properties that define words and used this knowledge to identify words they had never seen before. In addition, the pigeons were sensitive to the bigram frequencies of words (i.e., the common co-occurrence of certain letter pairs), the edit distance between nonwords and words, and the internal structure of words. Our findings demonstrate that visual systems organizationally distinct from the primate visual system can also be exapted or recycled to process the visual word form.
Collapse
|
26
|
Budgerigars and zebra finches differ in how they generalize in an artificial grammar learning experiment. Proc Natl Acad Sci U S A 2016; 113:E3977-84. [PMID: 27325756 DOI: 10.1073/pnas.1600483113] [Citation(s) in RCA: 46] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
The ability to abstract a regularity that underlies strings of sounds is a core mechanism of the language faculty but might not be specific to language learning or even to humans. It is unclear whether and to what extent nonhuman animals possess the ability to abstract regularities defining the relation among arbitrary auditory items in a string and to generalize this abstraction to strings of acoustically novel items. In this study we tested these abilities in a songbird (zebra finch) and a parrot species (budgerigar). Subjects were trained in a go/no-go design to discriminate between two sets of sound strings arranged in an XYX or an XXY structure. After this discrimination was acquired, each subject was tested with test strings that were structurally identical to the training strings but consisted of either new combinations of known elements or of novel elements belonging to other element categories. Both species learned to discriminate between the two stimulus sets. However, their responses to the test strings were strikingly different. Zebra finches categorized test stimuli with previously heard elements by the ordinal position that these elements occupied in the training strings, independent of string structure. In contrast, the budgerigars categorized both novel combinations of familiar elements as well as strings consisting of novel element types by their underlying structure. They thus abstracted the relation among items in the XYX and XXY structures, an ability similar to that shown by human infants and indicating a level of abstraction comparable to analogical reasoning.
Collapse
|
27
|
Kriengwatana B, Spierings MJ, ten Cate C. Auditory discrimination learning in zebra finches: effects of sex, early life conditions and stimulus characteristics. Anim Behav 2016. [DOI: 10.1016/j.anbehav.2016.03.028] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/18/2023]
|
28
|
Ono S, Okanoya K, Seki Y. Hierarchical emergence of sequence sensitivity in the songbird auditory forebrain. J Comp Physiol A Neuroethol Sens Neural Behav Physiol 2016; 202:163-83. [PMID: 26864094 DOI: 10.1007/s00359-016-1070-7] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2015] [Revised: 01/13/2016] [Accepted: 01/14/2016] [Indexed: 11/28/2022]
Abstract
Bengalese finches (Lonchura striata var. domestica) generate more complex sequences in their songs than zebra finches. Because of this, we chose this species to explore the signal processing of sound sequence in the primary auditory forebrain area, field L, and in a secondary area, the caudomedial nidopallium (NCM). We simultaneously recorded activity from multiple single units in urethane-anesthetized birds. We successfully replicated the results of a previous study in awake zebra finches examining stimulus-specific habituation of NCM neurons to conspecific songs. Then, we used an oddball paradigm and compared the neural response to deviant sounds that were presented infrequently, with the response to standard sounds, which were presented frequently. In a single sound oddball task, two different song elements were assigned for the deviant and standard sounds. The response bias to deviant elements was larger in NCM than in field L. In a triplet sequence oddball task, two triplet sequences containing elements ABC and ACB were assigned as the deviant and standard. Only neurons in NCM that displayed broad-shaped spike waveforms had sensitivity to the difference in element order. Our results suggest the hierarchical processing of complex sound sequences in the songbird auditory forebrain.
Collapse
Affiliation(s)
- Satoko Ono
- Department of Life Sciences, Graduate School of Arts and Sciences, The University of Tokyo, 3-8-1 Komaba, Meguro-ku, Tokyo, 153-8902, Japan.,ERATO, Okanoya Emotional Information Project, Japan Science and Technology Agency, 2-1 Hirosawa, Wako, Saitama, 351-0198, Japan.,Emotional Information Joint Research Laboratory, RIKEN BSI, 2-1 Hirosawa, Wako, Saitama, 351-0198, Japan
| | - Kazuo Okanoya
- Department of Life Sciences, Graduate School of Arts and Sciences, The University of Tokyo, 3-8-1 Komaba, Meguro-ku, Tokyo, 153-8902, Japan.,ERATO, Okanoya Emotional Information Project, Japan Science and Technology Agency, 2-1 Hirosawa, Wako, Saitama, 351-0198, Japan.,Emotional Information Joint Research Laboratory, RIKEN BSI, 2-1 Hirosawa, Wako, Saitama, 351-0198, Japan
| | - Yoshimasa Seki
- Department of Life Sciences, Graduate School of Arts and Sciences, The University of Tokyo, 3-8-1 Komaba, Meguro-ku, Tokyo, 153-8902, Japan. .,ERATO, Okanoya Emotional Information Project, Japan Science and Technology Agency, 2-1 Hirosawa, Wako, Saitama, 351-0198, Japan. .,Emotional Information Joint Research Laboratory, RIKEN BSI, 2-1 Hirosawa, Wako, Saitama, 351-0198, Japan. .,Faculty of Letters, Aichi University, 1-1 Machihata, Machihata-cho, Toyohashi, Aichi, 441-8522, Japan.
| |
Collapse
|
29
|
Cause and function in behavioural biology: A tribute to Jerry Hogan. Behav Processes 2015; 117:1-3. [DOI: 10.1016/j.beproc.2015.06.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
30
|
Spierings M, de Weger A, Ten Cate C. Pauses enhance chunk recognition in song element strings by zebra finches. Anim Cogn 2015; 18:867-74. [PMID: 25771964 PMCID: PMC4460290 DOI: 10.1007/s10071-015-0855-3] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2014] [Revised: 02/20/2015] [Accepted: 03/02/2015] [Indexed: 11/01/2022]
Abstract
When learning a language, it is crucial to know which syllables of a continuous sound string belong together as words. Human infants achieve this by attending to pauses between words or to the co-occurrence of syllables. It is not only humans that can segment a continuous string. Songbirds learning their song tend to copy 'chunks' from one or more tutors' songs and combine these into their own song. In the tutor songs, these chunks are often separated by pauses and a high co-occurrence of elements, suggesting that these features affect chunking and song learning. We examined experimentally whether the presence of pauses and element co-occurrence affect the ability of adult zebra finches to discriminate strings of song elements. Using a go/no-go design, two groups of birds were trained to discriminate between two strings. In one group (Pause-group), pauses were inserted between co-occurring element triplets in the strings, and in the other group (No-pause group), both strings were continuous. After making a correct discrimination, an individual proceeded to a reversal training using string segments. Segments were element triplets consistent in co-occurrence, triplets that were partly consistent in composition and triplets consisting of elements that did not co-occur in the strings. The Pause-group was faster in discriminating between the two strings. This group also responded differently to consistent triplets in the reversal training, compared to inconsistent triplets. The No-pause group did not differentiate among the triplet types. These results indicate that pauses in strings of song elements aid song discrimination and memorization of co-occurring element groups.
Collapse
Affiliation(s)
- Michelle Spierings
- Behavioural Biology, Institute of Biology Leiden (IBL), Leiden University, P.O. Box 9505, 2300 RA, Leiden, The Netherlands,
| | | | | |
Collapse
|