1
|
Tsunada J, Wang X, Eliades SJ. Multiple processes of vocal sensory-motor interaction in primate auditory cortex. Nat Commun 2024; 15:3093. [PMID: 38600118 PMCID: PMC11006904 DOI: 10.1038/s41467-024-47510-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2022] [Accepted: 04/02/2024] [Indexed: 04/12/2024] Open
Abstract
Sensory-motor interactions in the auditory system play an important role in vocal self-monitoring and control. These result from top-down corollary discharges, relaying predictions about vocal timing and acoustics. Recent evidence suggests such signals may be two distinct processes, one suppressing neural activity during vocalization and another enhancing sensitivity to sensory feedback, rather than a single mechanism. Single-neuron recordings have been unable to disambiguate due to overlap of motor signals with sensory inputs. Here, we sought to disentangle these processes in marmoset auditory cortex during production of multi-phrased 'twitter' vocalizations. Temporal responses revealed two timescales of vocal suppression: temporally-precise phasic suppression during phrases and sustained tonic suppression. Both components were present within individual neurons, however, phasic suppression presented broadly regardless of frequency tuning (gating), while tonic was selective for vocal frequencies and feedback (prediction). This suggests that auditory cortex is modulated by concurrent corollary discharges during vocalization, with different computational mechanisms.
Collapse
Affiliation(s)
- Joji Tsunada
- Auditory and Communication Systems Laboratory, Department of Otorhinolaryngology: Head and Neck Surgery, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA, USA
- Chinese Institute for Brain Research, Beijing, China
| | - Xiaoqin Wang
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Steven J Eliades
- Auditory and Communication Systems Laboratory, Department of Otorhinolaryngology: Head and Neck Surgery, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA, USA.
- Department of Head and Neck Surgery & Communication Sciences, Duke University School of Medicine, Durham, NC, USA.
| |
Collapse
|
2
|
Dureux A, Zanini A, Everling S. Mapping of facial and vocal processing in common marmosets with ultra-high field fMRI. Commun Biol 2024; 7:317. [PMID: 38480875 PMCID: PMC10937914 DOI: 10.1038/s42003-024-06002-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2023] [Accepted: 03/01/2024] [Indexed: 03/17/2024] Open
Abstract
Primate communication relies on multimodal cues, such as vision and audition, to facilitate the exchange of intentions, enable social interactions, avoid predators, and foster group cohesion during daily activities. Understanding the integration of facial and vocal signals is pivotal to comprehend social interaction. In this study, we acquire whole-brain ultra-high field (9.4 T) fMRI data from awake marmosets (Callithrix jacchus) to explore brain responses to unimodal and combined facial and vocal stimuli. Our findings reveal that the multisensory condition not only intensifies activations in the occipito-temporal face patches and auditory voice patches but also engages a more extensive network that includes additional parietal, prefrontal and cingulate areas, compared to the summed responses of the unimodal conditions. By uncovering the neural network underlying multisensory audiovisual integration in marmosets, this study highlights the efficiency and adaptability of the marmoset brain in processing facial and vocal social signals, providing significant insights into primate social communication.
Collapse
Affiliation(s)
- Audrey Dureux
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, N6A 5K8, Canada.
| | - Alessandro Zanini
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, N6A 5K8, Canada
| | - Stefan Everling
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, N6A 5K8, Canada
- Department of Physiology and Pharmacology, University of Western Ontario, London, ON, N6A 5K8, Canada
| |
Collapse
|
3
|
Banerjee A, Chen F, Druckmann S, Long MA. Temporal scaling of motor cortical dynamics reveals hierarchical control of vocal production. Nat Neurosci 2024; 27:527-535. [PMID: 38291282 DOI: 10.1038/s41593-023-01556-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2023] [Accepted: 12/13/2023] [Indexed: 02/01/2024]
Abstract
Neocortical activity is thought to mediate voluntary control over vocal production, but the underlying neural mechanisms remain unclear. In a highly vocal rodent, the male Alston's singing mouse, we investigate neural dynamics in the orofacial motor cortex (OMC), a structure critical for vocal behavior. We first describe neural activity that is modulated by component notes (~100 ms), probably representing sensory feedback. At longer timescales, however, OMC neurons exhibit diverse and often persistent premotor firing patterns that stretch or compress with song duration (~10 s). Using computational modeling, we demonstrate that such temporal scaling, acting through downstream motor production circuits, can enable vocal flexibility. These results provide a framework for studying hierarchical control circuits, a common design principle across many natural and artificial systems.
Collapse
Affiliation(s)
- Arkarup Banerjee
- NYU Neuroscience Institute, New York University Langone Health, New York, NY, USA.
- Department of Otolaryngology, New York University Langone Health, New York, NY, USA.
- Center for Neural Science, New York University, New York, NY, USA.
- Cold Spring Harbor Laboratory, Cold Spring Harbor, NY, USA.
| | - Feng Chen
- Department of Applied Physics, Stanford University, Stanford, CA, USA
| | - Shaul Druckmann
- Department of Neurobiology, Stanford University, Stanford, CA, USA
| | - Michael A Long
- NYU Neuroscience Institute, New York University Langone Health, New York, NY, USA.
- Department of Otolaryngology, New York University Langone Health, New York, NY, USA.
- Center for Neural Science, New York University, New York, NY, USA.
| |
Collapse
|
4
|
Khanna AR, Muñoz W, Kim YJ, Kfir Y, Paulk AC, Jamali M, Cai J, Mustroph ML, Caprara I, Hardstone R, Mejdell M, Meszéna D, Zuckerman A, Schweitzer J, Cash S, Williams ZM. Single-neuronal elements of speech production in humans. Nature 2024; 626:603-610. [PMID: 38297120 PMCID: PMC10866697 DOI: 10.1038/s41586-023-06982-w] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2023] [Accepted: 12/14/2023] [Indexed: 02/02/2024]
Abstract
Humans are capable of generating extraordinarily diverse articulatory movement combinations to produce meaningful speech. This ability to orchestrate specific phonetic sequences, and their syllabification and inflection over subsecond timescales allows us to produce thousands of word sounds and is a core component of language1,2. The fundamental cellular units and constructs by which we plan and produce words during speech, however, remain largely unknown. Here, using acute ultrahigh-density Neuropixels recordings capable of sampling across the cortical column in humans, we discover neurons in the language-dominant prefrontal cortex that encoded detailed information about the phonetic arrangement and composition of planned words during the production of natural speech. These neurons represented the specific order and structure of articulatory events before utterance and reflected the segmentation of phonetic sequences into distinct syllables. They also accurately predicted the phonetic, syllabic and morphological components of upcoming words and showed a temporally ordered dynamic. Collectively, we show how these mixtures of cells are broadly organized along the cortical column and how their activity patterns transition from articulation planning to production. We also demonstrate how these cells reliably track the detailed composition of consonant and vowel sounds during perception and how they distinguish processes specifically related to speaking from those related to listening. Together, these findings reveal a remarkably structured organization and encoding cascade of phonetic representations by prefrontal neurons in humans and demonstrate a cellular process that can support the production of speech.
Collapse
Affiliation(s)
- Arjun R Khanna
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - William Muñoz
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | | | - Yoav Kfir
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Angelique C Paulk
- Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Mohsen Jamali
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Jing Cai
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Martina L Mustroph
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Irene Caprara
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Richard Hardstone
- Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Mackenna Mejdell
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Domokos Meszéna
- Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | | | - Jeffrey Schweitzer
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Sydney Cash
- Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Ziv M Williams
- Department of Neurosurgery, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA.
- Harvard-MIT Division of Health Sciences and Technology, Boston, MA, USA.
- Harvard Medical School, Program in Neuroscience, Boston, MA, USA.
| |
Collapse
|
5
|
Tsunada J, Eliades SJ. Frontal-Auditory Cortical Interactions and Sensory Prediction During Vocal Production in Marmoset Monkeys. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.01.28.577656. [PMID: 38352422 PMCID: PMC10862695 DOI: 10.1101/2024.01.28.577656] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 02/21/2024]
Abstract
The control of speech and vocal production involves the calculation of error between the intended vocal output and the resulting auditory feedback. Consistent with this model, recent evidence has demonstrated that the auditory cortex is suppressed immediately before and during vocal production, yet is still sensitive to differences between vocal output and altered auditory feedback. This suppression has been suggested to be the result of top-down signals containing information about the intended vocal output, potentially originating from motor or other frontal cortical areas. However, whether such frontal areas are the source of suppressive and predictive signaling to the auditory cortex during vocalization is unknown. Here, we simultaneously recorded neural activity from both the auditory and frontal cortices of marmoset monkeys while they produced self-initiated vocalizations. We found increases in neural activity in both brain areas preceding the onset of vocal production, notably changes in both multi-unit activity and local field potential theta-band power. Connectivity analysis using Granger causality demonstrated that frontal cortex sends directed signaling to the auditory cortex during this pre-vocal period. Importantly, this pre-vocal activity predicted both vocalization-induced suppression of the auditory cortex as well as the acoustics of subsequent vocalizations. These results suggest that frontal cortical areas communicate with the auditory cortex preceding vocal production, with frontal-auditory signals that may reflect the transmission of sensory prediction information. This interaction between frontal and auditory cortices may contribute to mechanisms that calculate errors between intended and actual vocal outputs during vocal communication.
Collapse
Affiliation(s)
- Joji Tsunada
- Chinese Institute for Brain Research, Beijing, China
- Department of Veterinary Medicine, Faculty of Agriculture, Iwate University, Morioka, Iwate, Japan
| | - Steven J. Eliades
- Department of Head and Neck Surgery & Communication Sciences, Duke University School of Medicine, Durham, NC 27710, USA
| |
Collapse
|
6
|
Zhao L, Wang X. Frontal cortex activity during the production of diverse social communication calls in marmoset monkeys. Nat Commun 2023; 14:6634. [PMID: 37857618 PMCID: PMC10587070 DOI: 10.1038/s41467-023-42052-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2022] [Accepted: 09/28/2023] [Indexed: 10/21/2023] Open
Abstract
Vocal communication is essential for social behaviors in humans and non-human primates. While the frontal cortex is crucial to human speech production, its role in vocal production in non-human primates has long been questioned. It is unclear whether activities in the frontal cortex represent diverse vocal signals used in non-human primate communication. Here we studied single neuron activities and local field potentials (LFP) in the frontal cortex of male marmoset monkeys while the animal engaged in vocal exchanges with conspecifics in a social environment. We found that both single neuron activities and LFP were modulated by the production of each of the four major call types. Moreover, neural activities showed distinct patterns for different call types and theta-band LFP oscillations showed phase-locking to the phrases of twitter calls, suggesting a neural representation of vocalization features. Our results suggest important functions of the marmoset frontal cortex in supporting the production of diverse vocalizations in communication.
Collapse
Affiliation(s)
- Lingyun Zhao
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, The Johns Hopkins University School of Medicine, Baltimore, MD, 21205, USA.
- Department of Neurological Surgery, University of California, San Francisco, CA, 94158, USA.
| | - Xiaoqin Wang
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, The Johns Hopkins University School of Medicine, Baltimore, MD, 21205, USA.
| |
Collapse
|
7
|
Grijseels DM, Prendergast BJ, Gorman JC, Miller CT. The neurobiology of vocal communication in marmosets. Ann N Y Acad Sci 2023; 1528:13-28. [PMID: 37615212 PMCID: PMC10592205 DOI: 10.1111/nyas.15057] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/25/2023]
Abstract
An increasingly popular animal model for studying the neural basis of social behavior, cognition, and communication is the common marmoset (Callithrix jacchus). Interest in this New World primate across neuroscience is now being driven by their proclivity for prosociality across their repertoire, high volubility, and rapid development, as well as their amenability to naturalistic testing paradigms and freely moving neural recording and imaging technologies. The complement of these characteristics set marmosets up to be a powerful model of the primate social brain in the years to come. Here, we focus on vocal communication because it is the area that has both made the most progress and illustrates the prodigious potential of this species. We review the current state of the field with a focus on the various brain areas and networks involved in vocal perception and production, comparing the findings from marmosets to other animals, including humans.
Collapse
Affiliation(s)
- Dori M Grijseels
- Cortical Systems and Behavior Laboratory, University of California, San Diego, La Jolla, California, USA
| | - Brendan J Prendergast
- Cortical Systems and Behavior Laboratory, University of California, San Diego, La Jolla, California, USA
| | - Julia C Gorman
- Cortical Systems and Behavior Laboratory, University of California, San Diego, La Jolla, California, USA
- Neurosciences Graduate Program, University of California, San Diego, La Jolla, California, USA
| | - Cory T Miller
- Cortical Systems and Behavior Laboratory, University of California, San Diego, La Jolla, California, USA
- Neurosciences Graduate Program, University of California, San Diego, La Jolla, California, USA
| |
Collapse
|
8
|
Gan-Or B, London M. Cortical circuits modulate mouse social vocalizations. SCIENCE ADVANCES 2023; 9:eade6992. [PMID: 37774030 PMCID: PMC10541007 DOI: 10.1126/sciadv.ade6992] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/01/2022] [Accepted: 08/30/2023] [Indexed: 10/01/2023]
Abstract
Vocalizations provide a means of communication with high fidelity and information rate for many species. Diencephalon and brainstem neural circuits have been shown to control mouse vocal production; however, the role of cortical circuits in this process is debatable. Using electrical and optogenetic stimulation, we identified a cortical region in the anterior cingulate cortex in which stimulation elicits ultrasonic vocalizations. Moreover, fiber photometry showed an increase in Ca2+ dynamics preceding vocal initiation, whereas optogenetic suppression in this cortical area caused mice to emit fewer vocalizations. Last, electrophysiological recordings indicated a differential increase in neural activity in response to female social exposure dependent on vocal output. Together, these results indicate that the cortex is a key node in the neuronal circuits controlling vocal behavior in mice.
Collapse
Affiliation(s)
- Benjamin Gan-Or
- Edmond and Lily Safra Center for Brain Sciences and Alexander Silberman Institute of Life Sciences, The Hebrew University of Jerusalem, Jerusalem 91904, Israel
| | | |
Collapse
|
9
|
Risueno-Segovia C, Dohmen D, Gultekin YB, Pomberger T, Hage SR. Linguistic law-like compression strategies emerge to maximize coding efficiency in marmoset vocal communication. Proc Biol Sci 2023; 290:20231503. [PMID: 37752844 PMCID: PMC10523061 DOI: 10.1098/rspb.2023.1503] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Accepted: 08/29/2023] [Indexed: 09/28/2023] Open
Abstract
Human language follows statistical regularities or linguistic laws. For instance, Zipf's law of brevity states that the more frequently a word is used, the shorter it tends to be. All human languages adhere to this word structure. However, it is unclear whether Zipf's law emerged de novo in humans or whether it also exists in the non-linguistic vocal systems of our primate ancestors. Using a vocal conditioning paradigm, we examined the capacity of marmoset monkeys to efficiently encode vocalizations. We observed that marmosets adopted vocal compression strategies at three levels: (i) increasing call rate, (ii) decreasing call duration and (iii) increasing the proportion of short calls. Our results demonstrate that marmosets, when able to freely choose what to vocalize, exhibit vocal statistical regularities consistent with Zipf's law of brevity that go beyond their context-specific natural vocal behaviour. This suggests that linguistic laws emerged in non-linguistic vocal systems in the primate lineage.
Collapse
Affiliation(s)
- Cristina Risueno-Segovia
- Neurobiology of Social Communication, Department of Otolaryngology—Head and Neck Surgery, Hearing Research Centre, University of Tübingen, Medical Center, Elfriede-Aulhorn-Strasse 5, 72076 Tübingen, Germany
- Werner Reichardt Centre for Integrative Neuroscience, University of Tübingen, Otfried-Müller-Str. 25, 72076 Tübingen, Germany
- Graduate School of Neural & Behavioural Sciences - International Max Planck Research School, University of Tübingen, Österberg-Str. 3, 72074 Tübingen, Germany
| | - Deniz Dohmen
- Werner Reichardt Centre for Integrative Neuroscience, University of Tübingen, Otfried-Müller-Str. 25, 72076 Tübingen, Germany
- Graduate School of Neural & Behavioural Sciences - International Max Planck Research School, University of Tübingen, Österberg-Str. 3, 72074 Tübingen, Germany
| | - Yasemin B. Gultekin
- Neurobiology of Social Communication, Department of Otolaryngology—Head and Neck Surgery, Hearing Research Centre, University of Tübingen, Medical Center, Elfriede-Aulhorn-Strasse 5, 72076 Tübingen, Germany
- Werner Reichardt Centre for Integrative Neuroscience, University of Tübingen, Otfried-Müller-Str. 25, 72076 Tübingen, Germany
- Graduate School of Neural & Behavioural Sciences - International Max Planck Research School, University of Tübingen, Österberg-Str. 3, 72074 Tübingen, Germany
| | - Thomas Pomberger
- Neurobiology of Social Communication, Department of Otolaryngology—Head and Neck Surgery, Hearing Research Centre, University of Tübingen, Medical Center, Elfriede-Aulhorn-Strasse 5, 72076 Tübingen, Germany
- Werner Reichardt Centre for Integrative Neuroscience, University of Tübingen, Otfried-Müller-Str. 25, 72076 Tübingen, Germany
- Graduate School of Neural & Behavioural Sciences - International Max Planck Research School, University of Tübingen, Österberg-Str. 3, 72074 Tübingen, Germany
| | - Steffen R. Hage
- Neurobiology of Social Communication, Department of Otolaryngology—Head and Neck Surgery, Hearing Research Centre, University of Tübingen, Medical Center, Elfriede-Aulhorn-Strasse 5, 72076 Tübingen, Germany
- Werner Reichardt Centre for Integrative Neuroscience, University of Tübingen, Otfried-Müller-Str. 25, 72076 Tübingen, Germany
| |
Collapse
|
10
|
Jafari A, Dureux A, Zanini A, Menon RS, Gilbert KM, Everling S. A vocalization-processing network in marmosets. Cell Rep 2023; 42:112526. [PMID: 37195863 DOI: 10.1016/j.celrep.2023.112526] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2023] [Revised: 03/31/2023] [Accepted: 05/02/2023] [Indexed: 05/19/2023] Open
Abstract
Vocalizations play an important role in the daily life of primates and likely form the basis of human language. Functional imaging studies have demonstrated that listening to voices activates a fronto-temporal voice perception network in human participants. Here, we acquired whole-brain ultrahigh-field (9.4 T) fMRI in awake marmosets (Callithrix jacchus) and demonstrate that these small, highly vocal New World primates possess a similar fronto-temporal network, including subcortical regions, that is activated by the presentation of conspecific vocalizations. The findings suggest that the human voice perception network has evolved from an ancestral vocalization-processing network that predates the separation of New and Old World primates.
Collapse
Affiliation(s)
- Azadeh Jafari
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Audrey Dureux
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Alessandro Zanini
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Ravi S Menon
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Kyle M Gilbert
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada
| | - Stefan Everling
- Centre for Functional and Metabolic Mapping, Robarts Research Institute, University of Western Ontario, London, ON, Canada; Department of Physiology and Pharmacology, University of Western Ontario, London, ON, Canada.
| |
Collapse
|
11
|
Salles A, Neunuebel J. What do mammals have to say about the neurobiology of acoustic communication? MOLECULAR PSYCHOLOGY : BRAIN, BEHAVIOR, AND SOCIETY 2023; 2:5. [PMID: 38827277 PMCID: PMC11141777 DOI: 10.12688/molpsychol.17539.1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/04/2024]
Abstract
Auditory communication is crucial across taxa, including humans, because it enables individuals to convey information about threats, food sources, mating opportunities, and other social cues necessary for survival. Comparative approaches to auditory communication will help bridge gaps across taxa and facilitate our understanding of the neural mechanisms underlying this complex task. In this work, we briefly review the field of auditory communication processing and the classical champion animal, the songbird. In addition, we discuss other mammalian species that are advancing the field. In particular, we emphasize mice and bats, highlighting the characteristics that may inform how we think about communication processing.
Collapse
Affiliation(s)
- Angeles Salles
- Biological Sciences, University of Illinois Chicago, Chicago, Illinois, USA
| | - Joshua Neunuebel
- Psychological and Brain Sciences, University of Delaware, Newark, Delaware, USA
| |
Collapse
|
12
|
Chen C, Remington ED, Wang X. Sound localization acuity of the common marmoset (Callithrix jacchus). Hear Res 2023; 430:108722. [PMID: 36863289 DOI: 10.1016/j.heares.2023.108722] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/13/2022] [Revised: 02/03/2023] [Accepted: 02/10/2023] [Indexed: 02/14/2023]
Abstract
The common marmoset (Callithrix jacchus) is a small arboreal New World primate which has emerged as a promising model in auditory neuroscience. One potentially useful application of this model system is in the study of the neural mechanism underlying spatial hearing in primate species, as the marmosets need to localize sounds to orient their head to events of interest and identify their vocalizing conspecifics that are not visible. However, interpretation of neurophysiological data on sound localization requires an understanding of perceptual abilities, and the sound localization behavior of marmosets has not been well studied. The present experiment measured sound localization acuity using an operant conditioning procedure in which marmosets were trained to discriminate changes in sound location in the horizontal (azimuth) or vertical (elevation) dimension. Our results showed that the minimum audible angle (MAA) for horizontal and vertical discrimination was 13.17° and 12.53°, respectively, for 2 to 32 kHz Gaussian noise. Removing the monaural spectral cues tended to increase the horizontal localization acuity (11.31°). Marmosets have larger horizontal MAA (15.54°) in the rear than the front. Removing the high-frequency (> 26 kHz) region of the head-related transfer function (HRTF) affected vertical acuity mildly (15.76°), but removing the first notch (12-26 kHz) region of HRTF substantially reduced the vertical acuity (89.01°). In summary, our findings indicate that marmosets' spatial acuity is on par with other species of similar head size and field of best vision, and they do not appear to use monaural spectral cues for horizontal discrimination but rely heavily on first notch region of HRTF for vertical discrimination.
Collapse
Affiliation(s)
- Chenggang Chen
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, 720 Rutland Ave., Traylor 410, Baltimore, MD 21025, United States
| | - Evan D Remington
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, 720 Rutland Ave., Traylor 410, Baltimore, MD 21025, United States
| | - Xiaoqin Wang
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, 720 Rutland Ave., Traylor 410, Baltimore, MD 21025, United States.
| |
Collapse
|
13
|
Banerjee A, Chen F, Druckmann S, Long MA. Neural dynamics in the rodent motor cortex enables flexible control of vocal timing. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.01.23.525252. [PMID: 36747850 PMCID: PMC9900850 DOI: 10.1101/2023.01.23.525252] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 01/25/2023]
Abstract
Neocortical activity is thought to mediate voluntary control over vocal production, but the underlying neural mechanisms remain unclear. In a highly vocal rodent, the Alston's singing mouse, we investigate neural dynamics in the orofacial motor cortex (OMC), a structure critical for vocal behavior. We first describe neural activity that is modulated by component notes (approx. 100 ms), likely representing sensory feedback. At longer timescales, however, OMC neurons exhibit diverse and often persistent premotor firing patterns that stretch or compress with song duration (approx. 10 s). Using computational modeling, we demonstrate that such temporal scaling, acting via downstream motor production circuits, can enable vocal flexibility. These results provide a framework for studying hierarchical control circuits, a common design principle across many natural and artificial systems.
Collapse
Affiliation(s)
- Arkarup Banerjee
- NYU Neuroscience Institute, New York University Langone Health, New York, NY 10016, USA
- Department of Otolaryngology, New York University Langone Health, New York, NY 10016, USA
- Center for Neural Science, New York University, New York, NY 10003, USA
- Cold Spring Harbor Laboratory, Cold Spring Harbor, NY 11724, USA
| | - Feng Chen
- Department of Applied Physics, Stanford University, Stanford, CA 94305, USA
| | - Shaul Druckmann
- Department of Neuroscience, Stanford University, Stanford, CA 94304, USA
| | - Michael A Long
- NYU Neuroscience Institute, New York University Langone Health, New York, NY 10016, USA
- Department of Otolaryngology, New York University Langone Health, New York, NY 10016, USA
- Center for Neural Science, New York University, New York, NY 10003, USA
- Cold Spring Harbor Laboratory, Cold Spring Harbor, NY 11724, USA
| |
Collapse
|
14
|
Castellucci GA, Guenther FH, Long MA. A Theoretical Framework for Human and Nonhuman Vocal Interaction. Annu Rev Neurosci 2022; 45:295-316. [PMID: 35316612 PMCID: PMC9909589 DOI: 10.1146/annurev-neuro-111020-094807] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Vocal communication is a critical feature of social interaction across species; however, the relation between such behavior in humans and nonhumans remains unclear. To enable comparative investigation of this topic, we review the literature pertinent to interactive language use and identify the superset of cognitive operations involved in generating communicative action. We posit these functions comprise three intersecting multistep pathways: (a) the Content Pathway, which selects the movements constituting a response; (b) the Timing Pathway, which temporally structures responses; and (c) the Affect Pathway, which modulates response parameters according to internal state. These processing streams form the basis of the Convergent Pathways for Interaction framework, which provides a conceptual model for investigating the cognitive and neural computations underlying vocal communication across species.
Collapse
Affiliation(s)
- Gregg A. Castellucci
- NYU Neuroscience Institute and Department of Otolaryngology, New York University Langone Medical Center, New York, NY, USA
| | - Frank H. Guenther
- Departments of Speech, Language & Hearing Sciences and Biomedical Engineering, Boston University, Boston, MA, USA
| | - Michael A. Long
- NYU Neuroscience Institute and Department of Otolaryngology, New York University Langone Medical Center, New York, NY, USA
| |
Collapse
|
15
|
Echolocation-related reversal of information flow in a cortical vocalization network. Nat Commun 2022; 13:3642. [PMID: 35752629 PMCID: PMC9233670 DOI: 10.1038/s41467-022-31230-6] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2021] [Accepted: 05/30/2022] [Indexed: 11/09/2022] Open
Abstract
The mammalian frontal and auditory cortices are important for vocal behavior. Here, using local-field potential recordings, we demonstrate that the timing and spatial patterns of oscillations in the fronto-auditory network of vocalizing bats (Carollia perspicillata) predict the purpose of vocalization: echolocation or communication. Transfer entropy analyses revealed predominant top-down (frontal-to-auditory cortex) information flow during spontaneous activity and pre-vocal periods. The dynamics of information flow depend on the behavioral role of the vocalization and on the timing relative to vocal onset. We observed the emergence of predominant bottom-up (auditory-to-frontal) information transfer during the post-vocal period specific to echolocation pulse emission, leading to self-directed acoustic feedback. Electrical stimulation of frontal areas selectively enhanced responses to sounds in auditory cortex. These results reveal unique changes in information flow across sensory and frontal cortices, potentially driven by the purpose of the vocalization in a highly vocal mammalian model.
Collapse
|
16
|
Abstract
We examined the cortical control of a laryngeal muscle that is essential for vocalization in two monkey species that differ in their vocal motor skill. Our results suggest that enhancements in vocal skill are coupled to enlargements in the descending output from two premotor areas, ventral area 6 (area 6V) and the supplementary motor area (SMA). This result challenges the view that improvements in motor skills are due largely to changes in the output from the primary motor cortex. Marmosets display remarkable vocal motor abilities. Macaques do not. What is it about the marmoset brain that enables its skill in the vocal domain? We examined the cortical control of a laryngeal muscle that is essential for vocalization in both species. We found that, in both monkeys, multiple premotor areas in the frontal lobe along with the primary motor cortex (M1) are major sources of disynaptic drive to laryngeal motoneurons. Two of the premotor areas, ventral area 6 (area 6V) and the supplementary motor area (SMA), are a substantially larger source of descending output in marmosets. We propose that the enhanced vocal motor skills of marmosets are due, in part, to the expansion of descending output from these premotor areas.
Collapse
|
17
|
Palma M, Khoshnevis M, Lion M, Zenga C, Kefs S, Fallegger F, Schiavone G, Flandin IG, Lacour S, Yvert B. Chronic recording of cortical activity underlying vocalization in awake minipigs. J Neurosci Methods 2022; 366:109427. [PMID: 34852254 DOI: 10.1016/j.jneumeth.2021.109427] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2021] [Revised: 11/19/2021] [Accepted: 11/24/2021] [Indexed: 10/19/2022]
Abstract
BACKGROUND Investigating brain dynamics underlying vocal production in animals is a powerful way to inform on the neural bases of human speech. In particular, brain networks underlying vocal production in non-human primates show striking similarities with the human speech production network. However, despite increasing findings also in birds and more recently in rodents, the extent to which the primate vocal cortical network model generalizes to other non-primate mammals remains unclear. Especially, no domestic species has yet been proposed to investigate vocal brain activity using electrophysiological approaches. NEW METHOD In the present study, we introduce a novel experimental paradigm to identify the cortical dynamics underlying vocal production in behaving minipigs. A key problem to chronically implant cortical probes in pigs is the presence and growth of frontal sinuses extending caudally to the parietal bone and preventing safe access to neural structures with conventional craniotomy in adult animals. RESULTS Here we first show that implantations of soft ECoG grids can be done safely using conventional craniotomy in minipigs younger than 5 months, a period when sinuses are not yet well developed. Using wireless recordings in behaving animals, we further show activation of the motor and premotor cortex around the onset of vocal production of grunts, the most common vocalization of pigs. CONCLUSION These results suggest that minipigs, which are very loquacious and social animals, can be a good experimental large animal model to study the cortical bases of vocal production.
Collapse
Affiliation(s)
- Marie Palma
- Univ. Grenoble Alpes, Inserm, U1216, Grenoble Institut Neurosciences, 38000 Grenoble, France
| | - Mehrdad Khoshnevis
- Univ. Grenoble Alpes, Inserm, U1216, Grenoble Institut Neurosciences, 38000 Grenoble, France
| | - Marie Lion
- Univ. Grenoble Alpes, Inserm, U1216, Grenoble Institut Neurosciences, 38000 Grenoble, France
| | - Cyril Zenga
- Univ. Grenoble Alpes, Inserm, U1216, Grenoble Institut Neurosciences, 38000 Grenoble, France
| | - Samy Kefs
- CHU Grenoble Alpes, Clinique Universitaire de Cancérologie-Radiothérapie, 38000 Grenoble, France
| | - Florian Fallegger
- Ecole Polytechnique Fédérale de Lausanne (EPFL), Laboratory for Soft Bioelectronic Interfaces, Geneva, Switzerland
| | - Giuseppe Schiavone
- Ecole Polytechnique Fédérale de Lausanne (EPFL), Laboratory for Soft Bioelectronic Interfaces, Geneva, Switzerland
| | - Isabelle Gabelle Flandin
- CHU Grenoble Alpes, Clinique Universitaire de Cancérologie-Radiothérapie, 38000 Grenoble, France
| | - Stéphanie Lacour
- Ecole Polytechnique Fédérale de Lausanne (EPFL), Laboratory for Soft Bioelectronic Interfaces, Geneva, Switzerland
| | - Blaise Yvert
- Univ. Grenoble Alpes, Inserm, U1216, Grenoble Institut Neurosciences, 38000 Grenoble, France.
| |
Collapse
|
18
|
Bakola S, Burman KJ, Bednarek S, Chan JM, Jermakow N, Worthy KH, Majka P, Rosa MGP. Afferent Connections of Cytoarchitectural Area 6M and Surrounding Cortex in the Marmoset: Putative Homologues of the Supplementary and Pre-supplementary Motor Areas. Cereb Cortex 2021; 32:41-62. [PMID: 34255833 DOI: 10.1093/cercor/bhab193] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2021] [Revised: 06/07/2021] [Accepted: 06/07/2021] [Indexed: 01/02/2023] Open
Abstract
Cortical projections to the caudomedial frontal cortex were studied using retrograde tracers in marmosets. We tested the hypothesis that cytoarchitectural area 6M includes homologues of the supplementary and pre-supplementary motor areas (SMA and pre-SMA) of other primates. We found that, irrespective of the injection sites' location within 6M, over half of the labeled neurons were located in motor and premotor areas. Other connections originated in prefrontal area 8b, ventral anterior and posterior cingulate areas, somatosensory areas (3a and 1-2), and areas on the rostral aspect of the dorsal posterior parietal cortex. Although the origin of afferents was similar, injections in rostral 6M received higher percentages of prefrontal afferents, and fewer somatosensory afferents, compared to caudal injections, compatible with differentiation into SMA and pre-SMA. Injections rostral to 6M (area 8b) revealed a very different set of connections, with increased emphasis on prefrontal and posterior cingulate afferents, and fewer parietal afferents. The connections of 6M were also quantitatively different from those of the primary motor cortex, dorsal premotor areas, and cingulate motor area 24d. These results show that the cortical motor control circuit is conserved in simian primates, indicating that marmosets can be valuable models for studying movement planning and control.
Collapse
Affiliation(s)
- Sophia Bakola
- Department of Physiology and Neuroscience Program, Biomedicine Discovery Institute, Monash University, Clayton, VIC 3800, Australia.,Monash University Node, ARC Centre of Excellence for Integrative Brain Function, Monash University, Clayton, VIC 3800, Australia
| | - Kathleen J Burman
- Department of Physiology and Neuroscience Program, Biomedicine Discovery Institute, Monash University, Clayton, VIC 3800, Australia.,Monash University Node, ARC Centre of Excellence for Integrative Brain Function, Monash University, Clayton, VIC 3800, Australia
| | - Sylwia Bednarek
- Laboratory of Neuroinformatics, Nencki Institute of Experimental Biology of the Polish Academy of Sciences, 02-093 Warsaw, Poland
| | - Jonathan M Chan
- Department of Physiology and Neuroscience Program, Biomedicine Discovery Institute, Monash University, Clayton, VIC 3800, Australia.,Monash University Node, ARC Centre of Excellence for Integrative Brain Function, Monash University, Clayton, VIC 3800, Australia
| | - Natalia Jermakow
- Laboratory of Neuroinformatics, Nencki Institute of Experimental Biology of the Polish Academy of Sciences, 02-093 Warsaw, Poland
| | - Katrina H Worthy
- Department of Physiology and Neuroscience Program, Biomedicine Discovery Institute, Monash University, Clayton, VIC 3800, Australia
| | - Piotr Majka
- Monash University Node, ARC Centre of Excellence for Integrative Brain Function, Monash University, Clayton, VIC 3800, Australia.,Laboratory of Neuroinformatics, Nencki Institute of Experimental Biology of the Polish Academy of Sciences, 02-093 Warsaw, Poland
| | - Marcello G P Rosa
- Department of Physiology and Neuroscience Program, Biomedicine Discovery Institute, Monash University, Clayton, VIC 3800, Australia.,Monash University Node, ARC Centre of Excellence for Integrative Brain Function, Monash University, Clayton, VIC 3800, Australia
| |
Collapse
|
19
|
Gultekin YB, Hildebrand DGC, Hammerschmidt K, Hage SR. High plasticity in marmoset monkey vocal development from infancy to adulthood. SCIENCE ADVANCES 2021; 7:7/27/eabf2938. [PMID: 34193413 PMCID: PMC8245035 DOI: 10.1126/sciadv.abf2938] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/16/2020] [Accepted: 05/17/2021] [Indexed: 05/21/2023]
Abstract
The vocal behavior of human infants undergoes marked changes across their first year while becoming increasingly speech-like. Conversely, vocal development in nonhuman primates has been assumed to be largely predetermined and completed within the first postnatal months. Contradicting this assumption, we found a dichotomy between the development of call features and vocal sequences in marmoset monkeys, suggestive of a role for experience. While changes in call features were related to physical maturation, sequences of and transitions between calls remained flexible until adulthood. As in humans, marmoset vocal behavior developed in stages correlated with motor and social development stages. These findings are evidence for a prolonged phase of plasticity during marmoset vocal development, a crucial primate evolutionary preadaptation for the emergence of vocal learning and speech.
Collapse
Affiliation(s)
- Yasemin B Gultekin
- Neurobiology of Social Communication, Department of Otolaryngology - Head and Neck Surgery, Medical Center, University of Tübingen, 72076 Tübingen, Germany
- Werner Reichardt Centre for Integrative Neuroscience, University of Tübingen, 72076 Tübingen, Germany
| | - David G C Hildebrand
- Laboratory of Neural Systems, The Rockefeller University, New York, NY 10065, USA
| | - Kurt Hammerschmidt
- Cognitive Ethology Laboratory, German Primate Center, 37077 Göttingen, Germany
| | - Steffen R Hage
- Neurobiology of Social Communication, Department of Otolaryngology - Head and Neck Surgery, Medical Center, University of Tübingen, 72076 Tübingen, Germany.
- Werner Reichardt Centre for Integrative Neuroscience, University of Tübingen, 72076 Tübingen, Germany
| |
Collapse
|
20
|
Abstract
The common marmoset (Callithrix jacchus), a small New World primate, is receiving substantial attention in the neuroscience and biomedical science fields because its anatomical features, functional and behavioral characteristics, and reproductive features and its amenability to available genetic modification technologies make it an attractive experimental subject. In this review, I outline the progress of marmoset neuroscience research and summarize both the current status (opportunities and limitations) of and the future perspectives on the application of marmosets in neuroscience and disease modeling.
Collapse
Affiliation(s)
- Hideyuki Okano
- Department of Physiology, Keio University School of Medicine, Shinjuku-ku, Tokyo 160-8582, Japan; .,Laboratory for Marmoset Neural Architecture, RIKEN Center for Brain Science, Wako-shi, Saitama 351-0198, Japan
| |
Collapse
|
21
|
Nonverbal auditory communication - Evidence for integrated neural systems for voice signal production and perception. Prog Neurobiol 2020; 199:101948. [PMID: 33189782 DOI: 10.1016/j.pneurobio.2020.101948] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2020] [Revised: 10/12/2020] [Accepted: 11/04/2020] [Indexed: 12/24/2022]
Abstract
While humans have developed a sophisticated and unique system of verbal auditory communication, they also share a more common and evolutionarily important nonverbal channel of voice signaling with many other mammalian and vertebrate species. This nonverbal communication is mediated and modulated by the acoustic properties of a voice signal, and is a powerful - yet often neglected - means of sending and perceiving socially relevant information. From the viewpoint of dyadic (involving a sender and a signal receiver) voice signal communication, we discuss the integrated neural dynamics in primate nonverbal voice signal production and perception. Most previous neurobiological models of voice communication modelled these neural dynamics from the limited perspective of either voice production or perception, largely disregarding the neural and cognitive commonalities of both functions. Taking a dyadic perspective on nonverbal communication, however, it turns out that the neural systems for voice production and perception are surprisingly similar. Based on the interdependence of both production and perception functions in communication, we first propose a re-grouping of the neural mechanisms of communication into auditory, limbic, and paramotor systems, with special consideration for a subsidiary basal-ganglia-centered system. Second, we propose that the similarity in the neural systems involved in voice signal production and perception is the result of the co-evolution of nonverbal voice production and perception systems promoted by their strong interdependence in dyadic interactions.
Collapse
|
22
|
Common marmoset as a model primate for study of the motor control system. Curr Opin Neurobiol 2020; 64:103-110. [DOI: 10.1016/j.conb.2020.02.013] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2020] [Revised: 02/24/2020] [Accepted: 02/25/2020] [Indexed: 02/08/2023]
|
23
|
Dissociation of Unit Activity and Gamma Oscillations during Vocalization in Primate Auditory Cortex. J Neurosci 2020; 40:4158-4171. [PMID: 32295815 DOI: 10.1523/jneurosci.2749-19.2020] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2019] [Revised: 02/10/2020] [Accepted: 02/26/2020] [Indexed: 11/21/2022] Open
Abstract
Vocal production is a sensory-motor process in which auditory self-monitoring is used to ensure accurate communication. During vocal production, the auditory cortex of both humans and animals is suppressed, a phenomenon that plays an important role in self-monitoring and vocal motor control. However, the underlying neural mechanisms of this vocalization-induced suppression are unknown. γ-band oscillations (>25 Hz) have been implicated a variety of cortical functions and are thought to arise from activity of local inhibitory interneurons, but have not been studied during vocal production. We therefore examined γ-band activity in the auditory cortex of vocalizing marmoset monkeys, of either sex, and found that γ responses increased during vocal production. This increase in γ contrasts with simultaneously recorded suppression of single-unit and multiunit responses. Recorded vocal γ oscillations exhibited two separable components: a vocalization-specific nonsynchronized ("induced") response correlating with vocal suppression, and a synchronized ("evoked") response that was also present during passive sound playback. These results provide evidence for the role of cortical γ oscillations during inhibitory processing. Furthermore, the two distinct components of the γ response suggest possible mechanisms for vocalization-induced suppression, and may correspond to the sensory-motor integration of top-down and bottom-up inputs to the auditory cortex during vocal production.SIGNIFICANCE STATEMENT Vocal communication is important to both humans and animals. In order to ensure accurate information transmission, we must monitor our own vocal output. Surprisingly, spiking activity in the auditory cortex is suppressed during vocal production yet maintains sensitivity to the sound of our own voice ("feedback"). The mechanisms of this vocalization-induced suppression are unknown. Here we show that auditory cortical γ oscillations, which reflect interneuron activity, are actually increased during vocal production, the opposite response of that seen in spiking units. We discuss these results with proposed functions of γ activity during inhibitory sensory processing and coordination of different brain regions, suggesting a role in sensory-motor integration.
Collapse
|
24
|
Nieder A, Mooney R. The neurobiology of innate, volitional and learned vocalizations in mammals and birds. Philos Trans R Soc Lond B Biol Sci 2020; 375:20190054. [PMID: 31735150 PMCID: PMC6895551 DOI: 10.1098/rstb.2019.0054] [Citation(s) in RCA: 60] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/15/2019] [Indexed: 11/12/2022] Open
Abstract
Vocalization is an ancient vertebrate trait essential to many forms of communication, ranging from courtship calls to free verse. Vocalizations may be entirely innate and evoked by sexual cues or emotional state, as with many types of calls made in primates, rodents and birds; volitional, as with innate calls that, following extensive training, can be evoked by arbitrary sensory cues in non-human primates and corvid songbirds; or learned, acoustically flexible and complex, as with human speech and the courtship songs of oscine songbirds. This review compares and contrasts the neural mechanisms underlying innate, volitional and learned vocalizations, with an emphasis on functional studies in primates, rodents and songbirds. This comparison reveals both highly conserved and convergent mechanisms of vocal production in these different groups, despite their often vast phylogenetic separation. This similarity of central mechanisms for different forms of vocal production presents experimentalists with useful avenues for gaining detailed mechanistic insight into how vocalizations are employed for social and sexual signalling, and how they can be modified through experience to yield new vocal repertoires customized to the individual's social group. This article is part of the theme issue 'What can animal communication teach us about human language?'
Collapse
Affiliation(s)
- Andreas Nieder
- Animal Physiology Unit, Institute of Neurobiology, University Tübingen, Auf der Morgenstelle 28, 72076 Tübingen, Germany
| | - Richard Mooney
- Department of Neurobiology, Duke University School of Medicine, Durham, NC 27710, USA
| |
Collapse
|
25
|
Eliades SJ, Wang X. Corollary Discharge Mechanisms During Vocal Production in Marmoset Monkeys. BIOLOGICAL PSYCHIATRY. COGNITIVE NEUROSCIENCE AND NEUROIMAGING 2019; 4:805-812. [PMID: 31420219 PMCID: PMC6733626 DOI: 10.1016/j.bpsc.2019.06.008] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/22/2019] [Revised: 06/24/2019] [Accepted: 06/24/2019] [Indexed: 01/11/2023]
Abstract
Interactions between motor systems and sensory processing are ubiquitous throughout the animal kingdom and play an important role in many sensorimotor behaviors, including both human speech and animal vocalization. During vocal production, the auditory system plays important roles in both encoding feedback of produced sounds, allowing one to self-monitor for vocal errors, and simultaneously maintaining sensitivity to the outside acoustic environment. Supporting these roles is an efferent motor-to-sensory signal known as a corollary discharge. This review summarizes recent work on the role of such signaling during vocalization in the marmoset monkey, a nonhuman primate model of social vocal communication.
Collapse
Affiliation(s)
- Steven J. Eliades
- Auditory and Communication Systems Laboratory, Department of Otorhinolaryngology: Head and Neck Surgery, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA, U.S.A
| | - Xiaoqin Wang
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, U.S.A
| |
Collapse
|
26
|
Tramacere A, Wada K, Okanoya K, Iriki A, Ferrari PF. Auditory-Motor Matching in Vocal Recognition and Imitative Learning. Neuroscience 2019; 409:222-234. [PMID: 30742962 DOI: 10.1016/j.neuroscience.2019.01.056] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2018] [Revised: 01/10/2019] [Accepted: 01/28/2019] [Indexed: 10/27/2022]
Abstract
Songbirds possess mirror neurons (MNs) activating during the perception and execution of specific features of songs. These neurons are located in high vocal center (HVC), a premotor nucleus implicated in song perception, production and learning, making worth to inquire their properties and functions in vocal recognition and imitative learning. By integrating a body of brain and behavioral data, we discuss neurophysiology, anatomical, computational properties and possible functions of songbird MNs. We state that the neurophysiological properties of songbird MNs depends on sensorimotor regions that are outside the auditory neural system. Interestingly, songbirds MNs can be the result of the specific type of song representation possessed by some songbird species. At the functional level, we discuss whether songbird MNs are involved in others' song recognition, by dissecting the function of recognition in various different but possible overlapping processes: action-oriented perception, discriminative-oriented perception and identification of the signaler. We conclude that songbird MNs may be involved in recognizing other singer's vocalizations, while their role in imitative learning still require to solve how auditory feedback are used to correct own vocal performance to match the tutor song. Finally, we compare songbird and human mirror responses, hypothesizing a case of convergent evolution, and proposing new experimental directions.
Collapse
Affiliation(s)
- Antonella Tramacere
- Max Planck for the Science of Human History, DLCE Department, Jena, Kahlaische Str 10, 07745, Germany.
| | - Kazuhiro Wada
- Faculty of Science, Department of Biological Sciences, Hokkaido University, Kita-10 Nishi-8 Kita-ku, Sapporo 060-0810, Japan
| | - Kazuo Okanoya
- Department of Life Sciences, Graduate School of Arts and Sciences, The University of Tokyo, 153-8902 Tokyo, Japan
| | - Atsushi Iriki
- RIKEN Center for Brain Science, 351-0106 Saitama Prefecture, Wako, Hirosawa, Japan
| | - Pier F Ferrari
- Department of Medicine and Surgery, University of Parma, via Volturno, 43125, Italy; Institut des Sciences Cognitives Marc Jannerod, CNRS/Universite' Claude Bernard Lyon, 67 Pd Pinel 69675, Bron Cedex, France
| |
Collapse
|
27
|
Oller DK, Griebel U, Iyer SN, Jhang Y, Warlaumont AS, Dale R, Call J. Language Origins Viewed in Spontaneous and Interactive Vocal Rates of Human and Bonobo Infants. Front Psychol 2019; 10:729. [PMID: 31001176 PMCID: PMC6455048 DOI: 10.3389/fpsyg.2019.00729] [Citation(s) in RCA: 25] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2019] [Accepted: 03/15/2019] [Indexed: 01/18/2023] Open
Abstract
From the first months of life, human infants produce "protophones," speech-like, non-cry sounds, presumed absent, or only minimally present in other apes. But there have been no direct quantitative comparisons to support this presumption. In addition, by 2 months, human infants show sustained face-to-face interaction using protophones, a pattern thought also absent or very limited in other apes, but again, without quantitative comparison. Such comparison should provide evidence relevant to determining foundations of language, since substantially flexible vocalization, the inclination to explore vocalization, and the ability to interact socially by means of vocalization are foundations for language. Here we quantitatively compare data on vocalization rates in three captive bonobo (Pan paniscus) mother-infant pairs with various sources of data from our laboratories on human infant vocalization. Both humans and bonobos produced distress sounds (cries/screams) and laughter. The bonobo infants also produced sounds that were neither screams nor laughs and that showed acoustic similarities to the human protophones. These protophone-like sounds confirm that bonobo infants share with humans the capacity to produce vocalizations that appear foundational for language. Still, there were dramatic differences between the species in both quantity and function of the protophone and protophone-like sounds. The bonobo protophone-like sounds were far less frequent than the human protophones, and the human protophones were far less likely to be interpreted as complaints and more likely as vocal play. Moreover, we found extensive vocal interaction between human infants and mothers, but no vocal interaction in the bonobo mother-infant pairs-while bonobo mothers were physically responsive to their infants, we observed no case of a bonobo mother vocalization directed to her infant. Our cross-species comparison focuses on low- and moderate-arousal circumstances because we reason the roots of language entail vocalization not triggered by excitement, for example, during fighting or intense play. Language appears to be founded in flexible vocalization, used to regulate comfortable social interaction, to share variable affective states at various levels of arousal, and to explore vocalization itself.
Collapse
Affiliation(s)
- D. Kimbrough Oller
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States
- Konrad Lorenz Institute for Evolution and Cognition Research, Klosterneuburg, Austria
| | - Ulrike Griebel
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States
- Institute for Intelligent Systems, University of Memphis, Memphis, TN, United States
- Konrad Lorenz Institute for Evolution and Cognition Research, Klosterneuburg, Austria
| | - Suneeti Nathani Iyer
- Department of Communication Sciences and Special Education, University of Georgia, Athens, GA, United States
| | - Yuna Jhang
- Department of Speech-Language Pathology and Audiology, Chung Shan Medical University, Taichung, Taiwan
| | - Anne S. Warlaumont
- Department of Communication, University of California, Los Angeles, Los Angeles, CA, United States
| | - Rick Dale
- Department of Communication, University of California, Los Angeles, Los Angeles, CA, United States
| | - Josep Call
- School of Psychology and Neuroscience, University of St. Andrews, St. Andrews, United Kingdom
- Max Planck Institute for Evolutionary Anthropology, Leipzig, Germany
| |
Collapse
|
28
|
Okobi DE, Banerjee A, Matheson AMM, Phelps SM, Long MA. Motor cortical control of vocal interaction in neotropical singing mice. Science 2019; 363:983-988. [DOI: 10.1126/science.aau9480] [Citation(s) in RCA: 73] [Impact Index Per Article: 14.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2018] [Accepted: 01/23/2019] [Indexed: 12/25/2022]
Abstract
Like many adaptive behaviors, acoustic communication often requires rapid modification of motor output in response to sensory cues. However, little is known about the sensorimotor transformations that underlie such complex natural behaviors. In this study, we examine vocal exchanges in Alston’s singing mouse (Scotinomys teguina). We find that males modify singing behavior during social interactions on a subsecond time course that resembles both traditional sensorimotor tasks and conversational speech. We identify an orofacial motor cortical region and, via a series of perturbation experiments, demonstrate a hierarchical control of vocal production, with the motor cortex influencing the pacing of singing behavior on a moment-by-moment basis, enabling precise vocal interactions. These results suggest a systems-level framework for understanding the sensorimotor transformations that underlie natural social interactions.
Collapse
|
29
|
Hage SR. Auditory and audio-vocal responses of single neurons in the monkey ventral premotor cortex. Hear Res 2018; 366:82-89. [DOI: 10.1016/j.heares.2018.03.019] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/14/2017] [Revised: 03/15/2018] [Accepted: 03/19/2018] [Indexed: 02/07/2023]
|
30
|
|
31
|
Hage SR. Dual neural network model of speech and language evolution: new insights on flexibility of vocal production systems and involvement of frontal cortex. Curr Opin Behav Sci 2018. [DOI: 10.1016/j.cobeha.2018.02.010] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
32
|
Flinker A, Knight RT. Broca’s area in comprehension and production, insights from intracranial studies in humans. Curr Opin Behav Sci 2018. [DOI: 10.1016/j.cobeha.2018.04.012] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/08/2023]
|
33
|
A "voice patch" system in the primate brain for processing vocal information? Hear Res 2018; 366:65-74. [PMID: 29776691 DOI: 10.1016/j.heares.2018.04.010] [Citation(s) in RCA: 20] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/05/2018] [Revised: 04/14/2018] [Accepted: 04/25/2018] [Indexed: 12/13/2022]
Abstract
We review behavioural and neural evidence for the processing of information contained in conspecific vocalizations (CVs) in three primate species: humans, macaques and marmosets. We focus on abilities that are present and ecologically relevant in all three species: the detection and sensitivity to CVs; and the processing of identity cues in CVs. Current evidence, although fragmentary, supports the notion of a "voice patch system" in the primate brain analogous to the face patch system of visual cortex: a series of discrete, interconnected cortical areas supporting increasingly abstract representations of the vocal input. A central question concerns the degree to which the voice patch system is conserved in evolution. We outline challenges that arise and suggesting potential avenues for comparing the organization of the voice patch system across primate brains.
Collapse
|
34
|
Internal states and extrinsic factors both determine monkey vocal production. Proc Natl Acad Sci U S A 2018; 115:3978-3983. [PMID: 29581269 DOI: 10.1073/pnas.1722426115] [Citation(s) in RCA: 44] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
A key question for understanding speech evolution is whether or not the vocalizations of our closest living relatives-nonhuman primates-represent the precursors to speech. Some believe that primate vocalizations are not volitional but are instead inextricably linked to internal states like arousal and thus bear little resemblance to human speech. Others disagree and believe that since many primates can use their vocalizations strategically, this demonstrates a degree of voluntary vocal control. In the current study, we present a behavioral paradigm that reliably elicits different types of affiliative vocalizations from marmoset monkeys while measuring their heart rate fluctuations using noninvasive electromyography. By modulating both the physical distance between marmosets and the sensory information available to them, we find that arousal levels are linked, but not inextricably, to vocal production. Different arousal levels are, generally, associated with changes in vocal acoustics and the drive to produce different call types. However, in contexts where marmosets are interacting, the production of these different call types is also affected by extrinsic factors such as the timing of a conspecific's vocalization. These findings suggest that variability in vocal output as a function of context might reflect trade-offs between the drive to perpetuate vocal contact and conserving energy.
Collapse
|
35
|
Abstract
Vocalizations are a pervasive feature of nonhuman primate social life, yet we know surprisingly little about their function. We review studies supporting the hypothesis that many primate vocalizations function to facilitate social interactions by reducing uncertainty about the signaler's intentions and likely behavior. Such interactions help to establish and maintain the social bonds that increase reproductive success. Compared with humans, songbirds, and a few other mammals, primates have small vocal repertoires that show little acoustic modification during development. However, their ability to modify call usage is extensive and tuned to variation in the social context, including the historical relationship between caller and listener and the caller's assessment of how a listener is likely to respond. We suggest parallels between the decision to vocalize and neurophysiological studies of other, nonvocal social decisions between interacting monkeys. The selective factors driving the early stages of language evolution may have come from the need to make decisions about when and how to call within the context of social challenges.
Collapse
|
36
|
Development of stereotaxic recording system for awake marmosets (Callithrix jacchus). Neurosci Res 2018; 135:37-45. [PMID: 29317247 DOI: 10.1016/j.neures.2018.01.001] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2017] [Revised: 01/04/2018] [Accepted: 01/05/2018] [Indexed: 11/21/2022]
Abstract
The common marmoset has been proposed as a potential alternative to macaque monkey as a primate model for neuroscience and medical research. Here, we have newly developed a stereotaxic neuronal recording system for awake marmosets under the head-fixed condition by modifying that for macaque monkeys. Using this system, we recorded neuronal activity in the cerebral cortex of awake marmosets and successfully identified the primary motor cortex by intracortical microstimulation. Neuronal activities of deep brain structures, such as the basal ganglia, thalamus, and cerebellum, in awake marmosets were also successfully recorded referring to magnetic resonance images. Our system is suitable for functional mapping of the brain, since the large recording chamber allows access to arbitrary regions over almost the entire brain, and the recording electrode can be easily moved stereotaxically from one site to another. In addition, our system is desirable for neuronal recording during task performance to assess motor skills and cognitive function, as the marmoset sits in the marmoset chair and can freely use its hands. Moreover, our system can be used in combination with cutting-edge techniques, such as two-photon imaging and optogenetic manipulation. This recording system will contribute to boosting neuroscience and medical research using marmosets.
Collapse
|
37
|
Ferrari PF, Gerbella M, Coudé G, Rozzi S. Two different mirror neuron networks: The sensorimotor (hand) and limbic (face) pathways. Neuroscience 2017; 358:300-315. [PMID: 28687313 DOI: 10.1016/j.neuroscience.2017.06.052] [Citation(s) in RCA: 43] [Impact Index Per Article: 6.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2017] [Revised: 06/27/2017] [Accepted: 06/28/2017] [Indexed: 12/15/2022]
Abstract
The vast majority of functional studies investigating mirror neurons (MNs) explored their properties in relation to hand actions, and very few investigated how MNs respond to mouth actions or communicative gestures. Since hand and mouth MNs were recorded in two partially overlapping sectors of the ventral precentral cortex of the macaque monkey, there is a general assumption that they share a same neuroanatomical network, with the parietal cortex as a main source of visual information. In the current review, we challenge this perspective and describe the connectivity pattern of mouth MN sector. The mouth MNs F5/opercular region is connected with premotor, parietal areas mostly related to the somatosensory and motor representation of the face/mouth, and with area PrCO, involved in processing gustatory and somatosensory intraoral input. Unlike hand MNs, mouth MNs do not receive their visual input from parietal regions. Such information related to face/communicative behaviors could come from the ventrolateral prefrontal cortex. Further strong connections derive from limbic structures involved in encoding emotional facial expressions and motivational/reward processing. These brain structures include the anterior cingulate cortex, the anterior and mid-dorsal insula, orbitofrontal cortex and the basolateral amygdala. The mirror mechanism is therefore composed and supported by at least two different anatomical pathways: one is concerned with sensorimotor transformation in relation to reaching and hand grasping within the traditional parietal-premotor circuits; the second one is linked to the mouth/face motor control and is connected with limbic structures, involved in communication/emotions and reward processing.
Collapse
Affiliation(s)
- P F Ferrari
- Institut des Sciences Cognitives - Marc Jeannerod, CNRS/Université Claude Bernard Lyon, 67 Pinel, 69675 Bron Cedex, France; Dipartimento di Medicina e Chirurgia, Unità di Neuroscienze, 39 Volturno, 43125 Parma, Italy.
| | - M Gerbella
- Dipartimento di Medicina e Chirurgia, Unità di Neuroscienze, 39 Volturno, 43125 Parma, Italy; Istituto Italiano di Tecnologia (IIT), Center for Biomolecular Nanotechnologies, Lecce, Italy
| | - G Coudé
- Institut des Sciences Cognitives - Marc Jeannerod, CNRS/Université Claude Bernard Lyon, 67 Pinel, 69675 Bron Cedex, France
| | - S Rozzi
- Dipartimento di Medicina e Chirurgia, Unità di Neuroscienze, 39 Volturno, 43125 Parma, Italy
| |
Collapse
|
38
|
Abstract
Communication is an inherently interactive process that weaves together the fabric of both human and nonhuman primate societies. To investigate the properties of the primate brain during active social signaling, we recorded the responses of frontal cortex neurons as freely moving marmosets engaged in conversational exchanges with a visually occluded virtual marmoset. We found that small changes in firing rate (∼1 Hz) occurred across a broadly distributed population of frontal cortex neurons when marmosets heard a conspecific vocalization, and that these changes corresponded to subjects' likelihood of producing or withholding a vocal reply. Although the contributions of individual neurons were relatively small, large populations of neurons were able to clearly distinguish between these social contexts. Most significantly, this social context-dependent change in firing rate was evident even before subjects heard the vocalization, indicating that the probability of a conversational exchange was determined by the state of the frontal cortex at the time a vocalization was heard, and not by a decision driven by acoustic characteristics of the vocalization. We found that changes in neural activity scaled with the length of the conversation, with greater changes in firing rate evident for longer conversations. These data reveal specific and important facets of this neural activity that constrain its possible roles in active social signaling, and we hypothesize that the close coupling between frontal cortex activity and this natural, active primate social-signaling behavior facilitates social-monitoring mechanisms critical to conversational exchanges.SIGNIFICANCE STATEMENT We provide evidence for a novel pattern of neural activity in the frontal cortex of freely moving, naturally behaving, marmoset monkeys that may facilitate natural primate conversations. We discovered small (∼1 Hz), but reliable, changes in neural activity that occurred before marmosets even heard a conspecific vocalization that, as a population, almost perfectly predicted whether subjects would produce a vocalization in response. The change in the state of the frontal cortex persisted throughout the conversation and its magnitude scaled linearly with the length of the interaction. We hypothesize that this social context-dependent change in frontal cortex activity is supported by several mechanisms, such as social arousal and attention, and facilitates social monitoring critical for vocal coordination characteristic of human and nonhuman primate conversations.
Collapse
|
39
|
Sound-Making Actions Lead to Immediate Plastic Changes of Neuromagnetic Evoked Responses and Induced β-Band Oscillations during Perception. J Neurosci 2017; 37:5948-5959. [PMID: 28539421 DOI: 10.1523/jneurosci.3613-16.2017] [Citation(s) in RCA: 25] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2016] [Revised: 04/18/2017] [Accepted: 05/13/2017] [Indexed: 11/21/2022] Open
Abstract
Auditory and sensorimotor brain areas interact during the action-perception cycle of sound making. Neurophysiological evidence of a feedforward model of the action and its outcome has been associated with attenuation of the N1 wave of auditory evoked responses elicited by self-generated sounds, such as talking and singing or playing a musical instrument. Moreover, neural oscillations at β-band frequencies have been related to predicting the sound outcome after action initiation. We hypothesized that a newly learned action-perception association would immediately modify interpretation of the sound during subsequent listening. Nineteen healthy young adults (7 female, 12 male) participated in three magnetoencephalographic recordings while first passively listening to recorded sounds of a bell ringing, then actively striking the bell with a mallet, and then again listening to recorded sounds. Auditory cortex activity showed characteristic P1-N1-P2 waves. The N1 was attenuated during sound making, while P2 responses were unchanged. In contrast, P2 became larger when listening after sound making compared with the initial naive listening. The P2 increase occurred immediately, while in previous learning-by-listening studies P2 increases occurred on a later day. Also, reactivity of β-band oscillations, as well as θ coherence between auditory and sensorimotor cortices, was stronger in the second listening block. These changes were significantly larger than those observed in control participants (eight female, five male), who triggered recorded sounds by a key press. We propose that P2 characterizes familiarity with sound objects, whereas β-band oscillation signifies involvement of the action-perception cycle, and both measures objectively indicate functional neuroplasticity in auditory perceptual learning.SIGNIFICANCE STATEMENT While suppression of auditory responses to self-generated sounds is well known, it is not clear whether the learned action-sound association modifies subsequent perception. Our study demonstrated the immediate effects of sound-making experience on perception using magnetoencephalographic recordings, as reflected in the increased auditory evoked P2 wave, increased responsiveness of β oscillations, and enhanced connectivity between auditory and sensorimotor cortices. The importance of motor learning was underscored as the changes were much smaller in a control group using a key press to generate the sounds instead of learning to play the musical instrument. The results support the rapid integration of a feedforward model during perception and provide a neurophysiological basis for the application of music making in motor rehabilitation training.
Collapse
|
40
|
Teramoto Y, Takahashi DY, Holmes P, Ghazanfar AA. Vocal development in a Waddington landscape. eLife 2017; 6. [PMID: 28092262 PMCID: PMC5310845 DOI: 10.7554/elife.20782] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2016] [Accepted: 01/15/2017] [Indexed: 01/28/2023] Open
Abstract
Vocal development is the adaptive coordination of the vocal apparatus, muscles, the nervous system, and social interaction. Here, we use a quantitative framework based on optimal control theory and Waddington’s landscape metaphor to provide an integrated view of this process. With a biomechanical model of the marmoset monkey vocal apparatus and behavioral developmental data, we show that only the combination of the developing vocal tract, vocal apparatus muscles and nervous system can fully account for the patterns of vocal development. Together, these elements influence the shape of the monkeys’ vocal developmental landscape, tilting, rotating or shifting it in different ways. We can thus use this framework to make quantitative predictions regarding how interfering factors or experimental perturbations can change the landscape within a species, or to explain comparative differences in vocal development across species DOI:http://dx.doi.org/10.7554/eLife.20782.001 As infants develop they learn new behaviors and refine existing ones. For example, human infants progress from crying to babbling to producing speech-like sounds. A complex sequence of changes in muscles, the nervous system and in patterns of interactions with other individuals all contribute to these emerging behaviors. Despite this complexity, most studies of vocal development have only considered single factors in isolation. A study of speech development, for example, might examine how changes in the brain enable infants to imitate sounds. However, that same study will probably ignore how changes in the structure of the vocal cords, or in the behavior of the parents, also promote imitation. Young marmoset monkeys, like human infants, gradually develop from producing immature cries to adult-like calls. Teramoto, Takahashi et al. built a computational model of this process and compared the model to data from real animals. The first version of the model focused solely on how the marmosets’ vocal cords grow, and did not fully reproduce how adult-like calls emerge in real marmosets. Teramoto, Takahashi et al. therefore added factors to the model that simulate improvements in muscle control, learning in the nervous system and in the behavior of other animals. These findings show that, to reflect how adult-like calls emerge in real marmosets, the model needs to include all of these factors. The model developed by Teramoto, Takahashi et al. may also provide insights into why vocal learning and some other behaviors emerge in some species and not others. It may also be used to predict the consequences of disrupting individual processes in young animals at particular points in time and how such disruptions shape the way an animal develops on its way to adulthood. DOI:http://dx.doi.org/10.7554/eLife.20782.002
Collapse
Affiliation(s)
- Yayoi Teramoto
- Princeton Neuroscience Institute, Princeton University, Princeton, United States
| | - Daniel Y Takahashi
- Princeton Neuroscience Institute, Princeton University, Princeton, United States.,Department of Psychology, Princeton University, Princeton, United States
| | - Philip Holmes
- Princeton Neuroscience Institute, Princeton University, Princeton, United States.,Department of Mechanical and Aerospace Engineering and Program in Applied and Computational Mathematics, Princeton University, Princeton, United States
| | - Asif A Ghazanfar
- Princeton Neuroscience Institute, Princeton University, Princeton, United States.,Department of Psychology, Princeton University, Princeton, United States.,Department of Ecology and Evolutionary Biology, Princeton University, Princeton, United States
| |
Collapse
|
41
|
Limiting parental feedback disrupts vocal development in marmoset monkeys. Nat Commun 2017; 8:14046. [PMID: 28090084 PMCID: PMC5241798 DOI: 10.1038/ncomms14046] [Citation(s) in RCA: 69] [Impact Index Per Article: 9.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2016] [Accepted: 11/23/2016] [Indexed: 11/09/2022] Open
Abstract
Vocalizations of human infants undergo dramatic changes across the first year by becoming increasingly mature and speech-like. Human vocal development is partially dependent on learning by imitation through social feedback between infants and caregivers. Recent studies revealed similar developmental processes being influenced by parental feedback in marmoset monkeys for apparently innate vocalizations. Marmosets produce infant-specific vocalizations that disappear after the first postnatal months. However, it is yet unclear whether parental feedback is an obligate requirement for proper vocal development. Using quantitative measures to compare call parameters and vocal sequence structure we show that, in contrast to normally raised marmosets, marmosets that were separated from parents after the third postnatal month still produced infant-specific vocal behaviour at subadult stages. These findings suggest a significant role of social feedback on primate vocal development until the subadult stages and further show that marmoset monkeys are a compelling model system for early human vocal development.
Collapse
|