1
|
De Gregorio C, Maiolini M, Raimondi T, Carugati F, Miaretsoa L, Valente D, Torti V, Giacoma C, Ravignani A, Gamba M. Isochrony as ancestral condition to call and song in a primate. Ann N Y Acad Sci 2024; 1537:41-50. [PMID: 38925552 DOI: 10.1111/nyas.15151] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/28/2024]
Abstract
Animal songs differ from calls in function and structure, and have comparative and translational value, showing similarities to human music. Rhythm in music is often distributed in quantized classes of intervals known as rhythmic categories. These classes have been found in the songs of a few nonhuman species but never in their calls. Are rhythmic categories song-specific, as in human music, or can they transcend the song-call boundary? We analyze the vocal displays of one of the few mammals producing both songs and call sequences: Indri indri. We test whether rhythmic categories (a) are conserved across songs produced in different contexts, (b) exist in call sequences, and (c) differ between songs and call sequences. We show that rhythmic categories occur across vocal displays. Vocalization type and function modulate deployment of categories. We find isochrony (1:1 ratio, like the rhythm of a ticking clock) in all song types, but only advertisement songs show three rhythmic categories (1:1, 1:2, 2:1 ratios). Like songs, some call types are also isochronous. Isochrony is the backbone of most indri vocalizations, unlike human speech, where it is rare. In indri, isochrony underlies both songs and hierarchy-less call sequences and might be ancestral to both.
Collapse
Affiliation(s)
- Chiara De Gregorio
- Department of Life Sciences and Systems Biology, University of Torino, Turin, Italy
- Department of Psychology, University of Warwick, Coventry, UK
| | - Marco Maiolini
- Department of Life Sciences and Systems Biology, University of Torino, Turin, Italy
| | - Teresa Raimondi
- Department of Life Sciences and Systems Biology, University of Torino, Turin, Italy
- Department of Human Neurosciences, Sapienza University of Rome, Rome, Italy
| | - Filippo Carugati
- Department of Life Sciences and Systems Biology, University of Torino, Turin, Italy
| | - Longondraza Miaretsoa
- Groupe d'étude et de recherche sur les primates de Madagascar (GERP), Antananarivo, Madagascar
| | - Daria Valente
- Department of Life Sciences and Systems Biology, University of Torino, Turin, Italy
- Parco Natura Viva Garda Zoological Park (PNV), Verona, Italy
| | - Valeria Torti
- Department of Life Sciences and Systems Biology, University of Torino, Turin, Italy
| | - Cristina Giacoma
- Department of Life Sciences and Systems Biology, University of Torino, Turin, Italy
| | - Andrea Ravignani
- Department of Human Neurosciences, Sapienza University of Rome, Rome, Italy
- Comparative Bioacoustics Group, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus, Aalborg, Denmark
| | - Marco Gamba
- Department of Life Sciences and Systems Biology, University of Torino, Turin, Italy
| |
Collapse
|
2
|
Hersh TA, Ravignani A, Whitehead H. Cetaceans are the next frontier for vocal rhythm research. Proc Natl Acad Sci U S A 2024; 121:e2313093121. [PMID: 38814875 PMCID: PMC11194516 DOI: 10.1073/pnas.2313093121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/01/2024] Open
Abstract
While rhythm can facilitate and enhance many aspects of behavior, its evolutionary trajectory in vocal communication systems remains enigmatic. We can trace evolutionary processes by investigating rhythmic abilities in different species, but research to date has largely focused on songbirds and primates. We present evidence that cetaceans-whales, dolphins, and porpoises-are a missing piece of the puzzle for understanding why rhythm evolved in vocal communication systems. Cetaceans not only produce rhythmic vocalizations but also exhibit behaviors known or thought to play a role in the evolution of different features of rhythm. These behaviors include vocal learning abilities, advanced breathing control, sexually selected vocal displays, prolonged mother-infant bonds, and behavioral synchronization. The untapped comparative potential of cetaceans is further enhanced by high interspecific diversity, which generates natural ranges of vocal and social complexity for investigating various evolutionary hypotheses. We show that rhythm (particularly isochronous rhythm, when sounds are equally spaced in time) is prevalent in cetacean vocalizations but is used in different contexts by baleen and toothed whales. We also highlight key questions and research areas that will enhance understanding of vocal rhythms across taxa. By coupling an infraorder-level taxonomic assessment of vocal rhythm production with comparisons to other species, we illustrate how broadly comparative research can contribute to a more nuanced understanding of the prevalence, evolution, and possible functions of rhythm in animal communication.
Collapse
Affiliation(s)
- Taylor A. Hersh
- Marine Mammal Institute, Oregon State University, Newport, OR97365
- Comparative Bioacoustics Group, Max Planck Institute for Psycholinguistics, Nijmegen6525 XD, The Netherlands
- Department of Biology, Dalhousie University, HalifaxNS B3H 4R2, Canada
| | - Andrea Ravignani
- Comparative Bioacoustics Group, Max Planck Institute for Psycholinguistics, Nijmegen6525 XD, The Netherlands
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University, Aarhus8000, Denmark
- Department of Human Neurosciences, Sapienza University of Rome, Rome00185, Italy
| | - Hal Whitehead
- Department of Biology, Dalhousie University, HalifaxNS B3H 4R2, Canada
| |
Collapse
|
3
|
Sabharwal SR, Breaden M, Volpe G, Camurri A, Keller PE. Leadership dynamics in musical groups: Quantifying effects of musical structure on directionality of influence in concert performance videos. PLoS One 2024; 19:e0300663. [PMID: 38568939 PMCID: PMC10990194 DOI: 10.1371/journal.pone.0300663] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2023] [Accepted: 03/01/2024] [Indexed: 04/05/2024] Open
Abstract
Music ensemble performance provides an ecologically valid context for investigating leadership dynamics in small group interactions. Musical texture, specifically the relative salience of simultaneously sounding ensemble parts, is a feature that can potentially alter leadership dynamics by introducing hierarchical relationships between individual parts. The present study extended previous work on quantifying interpersonal coupling in musical ensembles by examining the relationship between musical texture and leader-follower relations, operationalised as directionality of influence between co-performers' body motion in concert video recordings. It was hypothesised that the directionality of influence, indexed by Granger Causality, would be greater for 'homophonic' textures with a clear distinction between melody and accompaniment parts than for 'polyphonic' textures with less distinction between melody and accompaniment. This hypothesis was tested by using pose estimation algorithms to track instrumentalists' body movements in a string quartet and a clarinet quintet, and then applying Granger Causality analysis to their head motion to estimate directional influence between instrumentalist pairs for sections of the pieces that varied in texture. It was found that Granger Causality values were generally higher (indicating greater directionality of influence) for homophonic than polyphonic textures. Furthermore, considering melody and accompaniment instrument roles revealed more evidence for the melody instrument influencing accompanying instruments than vice versa, plus a high degree of directionality among accompanying instruments, in homophonic textures. These observed patterns of directional information flow in co-performer body motion are consistent with changing leader-follower relations depending on hierarchical relations between ensemble parts in terms of the relative salience of melodic material in the musical texture. The finding that automatic pose estimation can detect modulations of leadership dynamics in standard video recordings under naturalistic performance conditions has implications for investigating interpersonal coordination in large-scale music video datasets representing different cultural traditions, and for exploring nonverbal communication in group activities more generally.
Collapse
Affiliation(s)
| | - Matthew Breaden
- MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, Australia
| | | | | | - Peter E. Keller
- MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, Australia
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus, Aarhus, Aalborg, Denmark
| |
Collapse
|
4
|
Osiecka AN, Fearey J, Ravignani A, Burchardt LS. Isochrony in barks of Cape fur seal ( Arctocephalus pusillus pusillus) pups and adults. Ecol Evol 2024; 14:e11085. [PMID: 38463637 PMCID: PMC10920323 DOI: 10.1002/ece3.11085] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Revised: 02/09/2024] [Accepted: 02/13/2024] [Indexed: 03/12/2024] Open
Abstract
Animal vocal communication often relies on call sequences. The temporal patterns of such sequences can be adjusted to other callers, follow complex rhythmic structures or exhibit a metronome-like pattern (i.e., isochronous). How regular are the temporal patterns in animal signals, and what influences their precision? If present, are rhythms already there early in ontogeny? Here, we describe an exploratory study of Cape fur seal (Arctocephalus pusillus pusillus) barks-a vocalisation type produced across many pinniped species in rhythmic, percussive bouts. This study is the first quantitative description of barking in Cape fur seal pups. We analysed the rhythmic structures of spontaneous barking bouts of pups and adult females from the breeding colony in Cape Cross, Namibia. Barks of adult females exhibited isochrony, that is they were produced at fairly regular points in time. Instead, intervals between pup barks were more variable, that is skipping a bark in the isochronous series occasionally. In both age classes, beat precision, that is how well the barks followed a perfect template, was worse when barking at higher rates. Differences could be explained by physiological factors, such as respiration or arousal. Whether, and how, isochrony develops in this species remains an open question. This study provides evidence towards a rhythmic production of barks in Cape fur seal pups and lays the groundwork for future studies to investigate the development of rhythm using multidimensional metrics.
Collapse
Affiliation(s)
- Anna N. Osiecka
- Department of Vertebrate Ecology and Zoology, Faculty of BiologyUniversity of GdańskGdańskPoland
- Behavioural Ecology Group, Section for Ecology and Evolution, Department of BiologyUniversity of CopenhagenCopenhagenDenmark
| | - Jack Fearey
- Sea Search Research and Conservation NPCCape TownSouth Africa
- Department of Statistical Sciences, Centre for Statistics in Ecology, Environment and ConservationUniversity of Cape TownCape TownWestern CapeSouth Africa
| | - Andrea Ravignani
- Comparative Bioacoustics GroupMax Planck Institute for PsycholinguisticsNijmegenThe Netherlands
- Center for Music in the Brain, Department of Clinical MedicineAarhus UniversityAarhus CDenmark
- Department of Human NeurosciencesSapienza University of RomeRomeItaly
| | - Lara S. Burchardt
- Comparative Bioacoustics GroupMax Planck Institute for PsycholinguisticsNijmegenThe Netherlands
- Leibniz‐Zentrum Allgemeine SprachwissenschaftBerlinGermany
| |
Collapse
|
5
|
Wiles TM, Mangalam M, Sommerfeld JH, Kim SK, Brink KJ, Charles AE, Grunkemeyer A, Kalaitzi Manifrenti M, Mastorakis S, Stergiou N, Likens AD. NONAN GaitPrint: An IMU gait database of healthy young adults. Sci Data 2023; 10:867. [PMID: 38052819 PMCID: PMC10698035 DOI: 10.1038/s41597-023-02704-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2023] [Accepted: 10/31/2023] [Indexed: 12/07/2023] Open
Abstract
An ongoing thrust of research focused on human gait pertains to identifying individuals based on gait patterns. However, no existing gait database supports modeling efforts to assess gait patterns unique to individuals. Hence, we introduce the Nonlinear Analysis Core (NONAN) GaitPrint database containing whole body kinematics and foot placement during self-paced overground walking on a 200-meter looping indoor track. Noraxon Ultium MotionTM inertial measurement unit (IMU) sensors sampled the motion of 35 healthy young adults (19-35 years old; 18 men and 17 women; mean ± 1 s.d. age: 24.6 ± 2.7 years; height: 1.73 ± 0.78 m; body mass: 72.44 ± 15.04 kg) over 18 4-min trials across two days. Continuous variables include acceleration, velocity, position, and the acceleration, velocity, position, orientation, and rotational velocity of each corresponding body segment, and the angle of each respective joint. The discrete variables include an exhaustive set of gait parameters derived from the spatiotemporal dynamics of foot placement. We technically validate our data using continuous relative phase, Lyapunov exponent, and Hurst exponent-nonlinear metrics quantifying different aspects of healthy human gait.
Collapse
Affiliation(s)
- Tyler M Wiles
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA
| | - Madhur Mangalam
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA
| | - Joel H Sommerfeld
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA
| | - Seung Kyeom Kim
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA
| | - Kolby J Brink
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA
| | - Anaelle Emeline Charles
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA
| | - Alli Grunkemeyer
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA
| | - Marilena Kalaitzi Manifrenti
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA
| | - Spyridon Mastorakis
- College of Information Science and Technology, University of Nebraska at Omaha, Omaha, NE, 68182, USA
| | - Nick Stergiou
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA
- Department of Physical Education and Sport Science, Aristotle University, Thessaloniki, Greece
| | - Aaron D Likens
- Division of Biomechanics and Research Development, Department of Biomechanics, and Center for Research in Human Movement Variability, University of Nebraska at Omaha, Omaha, NE, 68182, USA.
| |
Collapse
|
6
|
Alviar C, Fram N, Lense M. Quantifying tightness - Looseness of interactions with dynamical systems methods. A Comment on "Musical engagement as a duet of tight synchrony and loose interpretability" by T. C. Rabinowitch. Phys Life Rev 2023; 47:209-210. [PMID: 37949006 DOI: 10.1016/j.plrev.2023.10.031] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Accepted: 10/23/2023] [Indexed: 11/12/2023]
Affiliation(s)
- Camila Alviar
- Department of Otolaryngology - Head and Neck Surgery, Vanderbilt University Medical Center, 1408 17th Ave. S, Nashville, TN, 37212, United States of America.
| | - Noah Fram
- Department of Otolaryngology - Head and Neck Surgery, Vanderbilt University Medical Center, 1408 17th Ave. S, Nashville, TN, 37212, United States of America.
| | - Miriam Lense
- Department of Otolaryngology - Head and Neck Surgery, Vanderbilt University Medical Center, 1408 17th Ave. S, Nashville, TN, 37212, United States of America; Vanderbilt Kennedy Center, Vanderbilt University Medical Center, Nashville, TN, 37212, United States of America.
| |
Collapse
|
7
|
Izen SC, Cassano-Coleman RY, Piazza EA. Music as a window into real-world communication. Front Psychol 2023; 14:1012839. [PMID: 37496799 PMCID: PMC10368476 DOI: 10.3389/fpsyg.2023.1012839] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2022] [Accepted: 06/06/2023] [Indexed: 07/28/2023] Open
Abstract
Communication has been studied extensively in the context of speech and language. While speech is tremendously effective at transferring ideas between people, music is another communicative mode that has a unique power to bring people together and transmit a rich tapestry of emotions, through joint music-making and listening in a variety of everyday contexts. Research has begun to examine the behavioral and neural correlates of the joint action required for successful musical interactions, but it has yet to fully account for the rich, dynamic, multimodal nature of musical communication. We review the current literature in this area and propose that naturalistic musical paradigms will open up new ways to study communication more broadly.
Collapse
|
8
|
Berthet M, Coye C, Dezecache G, Kuhn J. Animal linguistics: a primer. Biol Rev Camb Philos Soc 2023; 98:81-98. [PMID: 36189714 PMCID: PMC10091714 DOI: 10.1111/brv.12897] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2022] [Revised: 08/10/2022] [Accepted: 08/12/2022] [Indexed: 01/12/2023]
Abstract
The evolution of language has been investigated by several research communities, including biologists and linguists, striving to highlight similar linguistic capacities across species. To date, however, no consensus exists on the linguistic capacities of non-human species. Major controversies remain on the use of linguistic terminology, analysis methods and behavioural data collection. The field of 'animal linguistics' has emerged to overcome these difficulties and attempt to reach uniform methods and terminology. This primer is a tutorial review of 'animal linguistics'. It describes the linguistic concepts of semantics, pragmatics and syntax, and proposes minimal criteria to be fulfilled to claim that a given species displays a particular linguistic capacity. Second, it reviews relevant methods successfully applied to the study of communication in animals and proposes a list of useful references to detect and overcome major pitfalls commonly observed in the collection of animal behaviour data. This primer represents a step towards mutual understanding and fruitful collaborations between linguists and biologists.
Collapse
Affiliation(s)
- Mélissa Berthet
- Institut Jean Nicod, Département d'études cognitives, ENS, EHESS, CNRS, PSL University, 75005, Paris, France.,Center for the Interdisciplinary Study of Language Evolution, University of Zürich, Affolternstrasse 56, 8050, Zurich, Switzerland.,Department of Comparative Language Science, University of Zürich, Affolternstrasse 56, 8050, Zurich, Switzerland
| | - Camille Coye
- Institut Jean Nicod, Département d'études cognitives, ENS, EHESS, CNRS, PSL University, 75005, Paris, France.,Center for Ecology and Conservation, Bioscience Department, University of Exeter, Penryn Campus, Penryn, TR10 9FE, UK
| | | | - Jeremy Kuhn
- Institut Jean Nicod, Département d'études cognitives, ENS, EHESS, CNRS, PSL University, 75005, Paris, France
| |
Collapse
|
9
|
Jermyn AS, Stevenson DJ, Levitin DJ. 1/f laws found in non-human music. Sci Rep 2023; 13:1324. [PMID: 36694022 PMCID: PMC9873655 DOI: 10.1038/s41598-023-28444-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Accepted: 01/18/2023] [Indexed: 01/25/2023] Open
Abstract
A compelling question at the intersection of physics, neuroscience, and evolutionary biology concerns the extent to which the brains of various species evolved to encode regularities of the physical world. It would be parsimonious and adaptive, for example, for brains to evolve an innate understanding of gravity and the laws of motion, and to be able to detect, auditorily, those patterns of noises that ambulatory creatures make when moving about the world. One such physical regularity of the world is fractal structure, generally characterized by power-law correlations or 1/f β spectral distributions. Such laws are found broadly in nature and human artifacts, from noise in physical systems, to coastline topography (e.g., the Richardson effect), to neuronal spike patterns. These distributions have also been found to hold for the rhythm and power spectral density of a wide array of human music, suggesting that human music incorporates regularities of the physical world that our species evolved to recognize and produce. Here we show for the first time that 1/fβ laws also govern the spectral density of a wide range of animal vocalizations (music), from songbirds, to whales, to howling wolves. We discovered this 1/fβ power-law distribution in the vocalizations within all of the 17 diverse species examined. Our results demonstrate that such power laws are prevalent in the animal kingdom, evidence that their brains have evolved a sensitivity to them as an aid in processing sensory features of the natural world.
Collapse
Affiliation(s)
- Adam S Jermyn
- Kavli Institute for Theoretical Physics, University of California at Santa Barbara, Santa Barbara, CA, 93106, USA
| | - David J Stevenson
- Division of Geology and Planetary Science, California Institute of Technology, Pasadena, CA, 91125, USA
| | - Daniel J Levitin
- Department of Psychology, School of Computer Science, and Schulich School of Music, McGill University, Montreal, QC, H3A 1B1, Canada.
| |
Collapse
|
10
|
Boorom O, Alviar C, Zhang Y, Muñoz VA, Kello CT, Lense MD. Child language and autism diagnosis impact hierarchical temporal structure of parent-child vocal interactions in early childhood. Autism Res 2022; 15:2099-2111. [PMID: 36056678 PMCID: PMC9995224 DOI: 10.1002/aur.2804] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2022] [Accepted: 08/15/2022] [Indexed: 12/15/2022]
Abstract
Timing is critical to successful social interactions. The temporal structure of dyadic vocal interactions emerges from the rhythm, timing, and frequency of each individuals' vocalizations and reflects how the dyad dynamically organizes and adapts during an interaction. This study investigated the temporal structure of vocal interactions longitudinally in parent-child dyads of typically developing (TD) infants (n = 49; 9-18 months; 48% male) and toddlers with ASD (n = 23; 27.2 ± 5.0 months; 91.3% male) to identify how developing language and social skills impact the temporal dynamics of the interaction. Acoustic hierarchical temporal structure (HTS), a measure of the nested clustering of acoustic events across multiple timescales, was measured in free play interactions using Allan Factor. HTS reflects a signal's temporal complexity and variability, with greater HTS indicating reduced flexibility of the dyadic system. Child expressive language significantly predicted HTS (ß = -0.2) longitudinally across TD infants, with greater dyadic HTS associated with lower child language skills. ASD dyads exhibited greater HTS (i.e., more rigid temporal structure) than nonverbal matched (d = 0.41) and expressive language matched TD dyads (d = 0.28). Increased HTS in ASD dyads occurred at timescales >1 s, suggesting greater structuring of pragmatic aspects of interaction. Results provide a new window into how language development and social reciprocity serve as constraints to shape parent-child interaction dynamics and showcase a novel automated approach to characterizing vocal interactions across multiple timescales during early childhood.
Collapse
Affiliation(s)
- Olivia Boorom
- Department of Speech-Language-Hearing: Sciences and Disorders, University of Kansas, Lawrence, KS, USA
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN, USA
| | - Camila Alviar
- Department of Otolaryngology - Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, TN, USA
- Cognitive and Information Sciences, University of California, Merced, Merced, CA, USA
| | - Yumeng Zhang
- Department of Otolaryngology - Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, TN, USA
- Department of Medicine, Health, and Society, Vanderbilt University, Nashville, TN, USA
| | - Valerie A. Muñoz
- Department of Otolaryngology - Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, TN, USA
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, USA
| | - Christopher T. Kello
- Cognitive and Information Sciences, University of California, Merced, Merced, CA, USA
| | - Miriam D. Lense
- Department of Otolaryngology - Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, TN, USA
- Vanderbilt Kennedy Center, Vanderbilt University Medical Center, Nashville, TN, USA
| |
Collapse
|
11
|
Daikoku T, Goswami U. Hierarchical amplitude modulation structures and rhythm patterns: Comparing Western musical genres, song, and nature sounds to Babytalk. PLoS One 2022; 17:e0275631. [PMID: 36240225 PMCID: PMC9565671 DOI: 10.1371/journal.pone.0275631] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2021] [Accepted: 09/20/2022] [Indexed: 11/19/2022] Open
Abstract
Statistical learning of physical stimulus characteristics is important for the development of cognitive systems like language and music. Rhythm patterns are a core component of both systems, and rhythm is key to language acquisition by infants. Accordingly, the physical stimulus characteristics that yield speech rhythm in "Babytalk" may also describe the hierarchical rhythmic relationships that characterize human music and song. Computational modelling of the amplitude envelope of "Babytalk" (infant-directed speech, IDS) using a demodulation approach (Spectral-Amplitude Modulation Phase Hierarchy model, S-AMPH) can describe these characteristics. S-AMPH modelling of Babytalk has shown previously that bands of amplitude modulations (AMs) at different temporal rates and their phase relations help to create its structured inherent rhythms. Additionally, S-AMPH modelling of children's nursery rhymes shows that different rhythm patterns (trochaic, iambic, dactylic) depend on the phase relations between AM bands centred on ~2 Hz and ~5 Hz. The importance of these AM phase relations was confirmed via a second demodulation approach (PAD, Probabilistic Amplitude Demodulation). Here we apply both S-AMPH and PAD to demodulate the amplitude envelopes of Western musical genres and songs. Quasi-rhythmic and non-human sounds found in nature (birdsong, rain, wind) were utilized for control analyses. We expected that the physical stimulus characteristics in human music and song from an AM perspective would match those of IDS. Given prior speech-based analyses, we also expected that AM cycles derived from the modelling may identify musical units like crotchets, quavers and demi-quavers. Both models revealed an hierarchically-nested AM modulation structure for music and song, but not nature sounds. This AM modulation structure for music and song matched IDS. Both models also generated systematic AM cycles yielding musical units like crotchets and quavers. Both music and language are created by humans and shaped by culture. Acoustic rhythm in IDS and music appears to depend on many of the same physical characteristics, facilitating learning.
Collapse
Affiliation(s)
- Tatsuya Daikoku
- Centre for Neuroscience in Education, University of Cambridge, Cambridge, United Kingdom
- International Research Center for Neurointelligence, The University of Tokyo, Bunkyo City, Tokyo, Japan
- Center for Brain, Mind and KANSEI Sciences Research, Hiroshima University, Hiroshima, Japan
- * E-mail:
| | - Usha Goswami
- Centre for Neuroscience in Education, University of Cambridge, Cambridge, United Kingdom
| |
Collapse
|
12
|
Xing J, Sainburg T, Taylor H, Gentner TQ. Syntactic modulation of rhythm in Australian pied butcherbird song. ROYAL SOCIETY OPEN SCIENCE 2022; 9:220704. [PMID: 36177196 PMCID: PMC9515642 DOI: 10.1098/rsos.220704] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/27/2022] [Accepted: 09/05/2022] [Indexed: 05/04/2023]
Abstract
The acoustic structure of birdsong is spectrally and temporally complex. Temporal complexity is often investigated in a syntactic framework focusing on the statistical features of symbolic song sequences. Alternatively, temporal patterns can be investigated in a rhythmic framework that focuses on the relative timing between song elements. Here, we investigate the merits of combining both frameworks by integrating syntactic and rhythmic analyses of Australian pied butcherbird (Cracticus nigrogularis) songs, which exhibit organized syntax and diverse rhythms. We show that rhythms of the pied butcherbird song bouts in our sample are categorically organized and predictable by the song's first-order sequential syntax. These song rhythms remain categorically distributed and strongly associated with the first-order sequential syntax even after controlling for variance in note length, suggesting that the silent intervals between notes induce a rhythmic structure on note sequences. We discuss the implication of syntactic-rhythmic relations as a relevant feature of song complexity with respect to signals such as human speech and music, and advocate for a broader conception of song complexity that takes into account syntax, rhythm, and their interaction with other acoustic and perceptual features.
Collapse
Affiliation(s)
- Jeffrey Xing
- Department of Psychology, University of California San Diego, La Jolla, CA, USA
| | - Tim Sainburg
- Department of Psychology, University of California San Diego, La Jolla, CA, USA
| | - Hollis Taylor
- Sydney Conservatorium of Music, University of Sydney, Sydney, New South Wales, Australia
| | - Timothy Q. Gentner
- Department of Psychology, University of California San Diego, La Jolla, CA, USA
- Neurobiology Section, Division of Biological Sciences, University of California San Diego, La Jolla, CA, USA
- Kavli Institute for Brain and Mind, University of California San Diego, La Jolla, CA, USA
| |
Collapse
|
13
|
Xing J, Sainburg T, Taylor H, Gentner TQ. Syntactic modulation of rhythm in Australian pied butcherbird song. ROYAL SOCIETY OPEN SCIENCE 2022; 9:220704. [PMID: 36177196 DOI: 10.6084/m9.figshare.c.6197494] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Subscribe] [Scholar Register] [Received: 05/27/2022] [Accepted: 09/05/2022] [Indexed: 05/21/2023]
Abstract
The acoustic structure of birdsong is spectrally and temporally complex. Temporal complexity is often investigated in a syntactic framework focusing on the statistical features of symbolic song sequences. Alternatively, temporal patterns can be investigated in a rhythmic framework that focuses on the relative timing between song elements. Here, we investigate the merits of combining both frameworks by integrating syntactic and rhythmic analyses of Australian pied butcherbird (Cracticus nigrogularis) songs, which exhibit organized syntax and diverse rhythms. We show that rhythms of the pied butcherbird song bouts in our sample are categorically organized and predictable by the song's first-order sequential syntax. These song rhythms remain categorically distributed and strongly associated with the first-order sequential syntax even after controlling for variance in note length, suggesting that the silent intervals between notes induce a rhythmic structure on note sequences. We discuss the implication of syntactic-rhythmic relations as a relevant feature of song complexity with respect to signals such as human speech and music, and advocate for a broader conception of song complexity that takes into account syntax, rhythm, and their interaction with other acoustic and perceptual features.
Collapse
Affiliation(s)
- Jeffrey Xing
- Department of Psychology, University of California San Diego, La Jolla, CA, USA
| | - Tim Sainburg
- Department of Psychology, University of California San Diego, La Jolla, CA, USA
| | - Hollis Taylor
- Sydney Conservatorium of Music, University of Sydney, Sydney, New South Wales, Australia
| | - Timothy Q Gentner
- Department of Psychology, University of California San Diego, La Jolla, CA, USA
- Neurobiology Section, Division of Biological Sciences, University of California San Diego, La Jolla, CA, USA
- Kavli Institute for Brain and Mind, University of California San Diego, La Jolla, CA, USA
| |
Collapse
|
14
|
Mangalam M, Kelty-Stephen DG. Ergodic descriptors of non-ergodic stochastic processes. J R Soc Interface 2022; 19:20220095. [PMID: 35414215 PMCID: PMC9006033 DOI: 10.1098/rsif.2022.0095] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
The stochastic processes underlying the growth and stability of biological and psychological systems reveal themselves when far-from-equilibrium. Far-from-equilibrium, non-ergodicity reigns. Non-ergodicity implies that the average outcome for a group/ensemble (i.e. of representative organisms/minds) is not necessarily a reliable estimate of the average outcome for an individual over time. However, the scientific interest in causal inference suggests that we somehow aim at stable estimates of the cause that will generalize to new individuals in the long run. Therefore, the valid analysis must extract an ergodic stationary measure from fluctuating physiological data. So the challenge is to extract statistical estimates that may describe or quantify some of this non-ergodicity (i.e. of the raw measured data) without themselves (i.e. the estimates) being non-ergodic. We show that traditional linear statistics such as the standard deviation, coefficient of variation and root mean square can break ergodicity. Time series of statistics addressing sequential structure and its potential nonlinearity: fractality and multi-fractality, change in a time-independent way and fulfil the ergodic assumption. Complementing traditional linear indices with fractal and multi-fractal indices would empower the study of stochastic far-from-equilibrium biological and psychological dynamics.
Collapse
Affiliation(s)
- Madhur Mangalam
- Department of Physical Therapy, Movement and Rehabilitation Sciences, Northeastern University, Boston, MA, USA
| | | |
Collapse
|
15
|
Farrera A, Ramos-Fernández G. Collective Rhythm as an Emergent Property During Human Social Coordination. Front Psychol 2022; 12:772262. [PMID: 35222144 PMCID: PMC8868940 DOI: 10.3389/fpsyg.2021.772262] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2021] [Accepted: 11/30/2021] [Indexed: 11/23/2022] Open
Abstract
The literature on social interactions has shown that participants coordinate not only at the behavioral but also at the physiological and neural levels, and that this coordination gives a temporal structure to the individual and social dynamics. However, it has not been fully explored whether such temporal patterns emerge during interpersonal coordination beyond dyads, whether this phenomenon arises from complex cognitive mechanisms or from relatively simple rules of behavior, or which are the sociocultural processes that underlie this phenomenon. We review the evidence for the existence of group-level rhythmic patterns that result from social interactions and argue that the complexity of group dynamics can lead to temporal regularities that cannot be predicted from the individual periodicities: an emergent collective rhythm. Moreover, we use this interpretation of the literature to discuss how taking into account the sociocultural niche in which individuals develop can help explain the seemingly divergent results that have been reported on the social influences and consequences of interpersonal coordination. We make recommendations on further research to test these arguments and their relationship to the feeling of belonging and assimilation experienced during group dynamics.
Collapse
Affiliation(s)
- Arodi Farrera
- Mathematical Modeling of Social Systems Department, Institute for Research on Applied Mathematics and Systems, National Autonomous University of Mexico, Mexico City, Mexico
| | - Gabriel Ramos-Fernández
- Mathematical Modeling of Social Systems Department, Institute for Research on Applied Mathematics and Systems, National Autonomous University of Mexico, Mexico City, Mexico
| |
Collapse
|
16
|
Warlaumont AS, Sobowale K, Fausey CM. Daylong Mobile Audio Recordings Reveal Multitimescale Dynamics in Infants' Vocal Productions and Auditory Experiences. CURRENT DIRECTIONS IN PSYCHOLOGICAL SCIENCE 2022; 31:12-19. [PMID: 35707791 PMCID: PMC9197087 DOI: 10.1177/09637214211058166] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/20/2023]
Abstract
The sounds of human infancy-baby babbling, adult talking, lullaby singing, and more-fluctuate over time. Infant-friendly wearable audio recorders can now capture very large quantities of these sounds throughout infants' everyday lives at home. Here, we review recent discoveries about how infants' soundscapes are organized over the course of a day based on analyses designed to detect patterns at multiple timescales. Analyses of infants' day-long audio have revealed that everyday vocalizations are clustered hierarchically in time, vocal explorations are consistent with foraging dynamics, and musical tunes are distributed such that some are much more available than others. This approach focusing on the multi-scale distributions of sounds heard and produced by infants provides new, fundamental insights on human communication development from a complex systems perspective.
Collapse
Affiliation(s)
| | - Kunmi Sobowale
- Department of Psychiatry and Biobehavioral Sciences, University of California, Los Angeles
| | | |
Collapse
|
17
|
Proksch S, Reeves M, Spivey M, Balasubramaniam R. Coordination dynamics of multi-agent interaction in a musical ensemble. Sci Rep 2022; 12:421. [PMID: 35013620 PMCID: PMC8748883 DOI: 10.1038/s41598-021-04463-6] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2021] [Accepted: 12/22/2021] [Indexed: 11/21/2022] Open
Abstract
Humans interact with other humans at a variety of timescales and in a variety of social contexts. We exhibit patterns of coordination that may differ depending on whether we are genuinely interacting as part of a coordinated group of individuals vs merely co-existing within the same physical space. Moreover, the local coordination dynamics of an interacting pair of individuals in an otherwise non-interacting group may spread, propagating change in the global coordination dynamics and interaction of an entire crowd. Dynamical systems analyses, such as Recurrence Quantification Analysis (RQA), can shed light on some of the underlying coordination dynamics of multi-agent human interaction. We used RQA to examine the coordination dynamics of a performance of "Welcome to the Imagination World", composed for wind orchestra. This performance enacts a real-life simulation of the transition from uncoordinated, non-interacting individuals to a coordinated, interacting multi-agent group. Unlike previous studies of social interaction in musical performance which rely on different aspects of video and/or acoustic data recorded from each individual, this project analyzes group-level coordination patterns solely from the group-level acoustic data of an audio recording of the performance. Recurrence and stability measures extracted from the audio recording increased when musicians coordinated as an interacting group. Variability in these measures also increased, indicating that the interacting ensemble of musicians were able to explore a greater variety of behavior than when they performed as non-interacting individuals. As an orchestrated (non-emergent) example of coordination, we believe these analyses provide an indication of approximate expected distributions for recurrence patterns that may be measurable before and after truly emergent coordination.
Collapse
Affiliation(s)
- Shannon Proksch
- Cognitive and Information Sciences, University of California-Merced, Merced, USA.
| | - Majerle Reeves
- Applied Mathematics, University of California-Merced, Merced, USA
| | - Michael Spivey
- Cognitive and Information Sciences, University of California-Merced, Merced, USA
| | | |
Collapse
|
18
|
Lee J, Rothschild G. Encoding of acquired sound-sequence salience by auditory cortical offset responses. Cell Rep 2021; 37:109927. [PMID: 34731615 DOI: 10.1016/j.celrep.2021.109927] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2021] [Revised: 08/19/2021] [Accepted: 10/12/2021] [Indexed: 11/25/2022] Open
Abstract
Behaviorally relevant sounds are often composed of distinct acoustic units organized into specific temporal sequences. The meaning of such sound sequences can therefore be fully recognized only when they have terminated. However, the neural mechanisms underlying the perception of sound sequences remain unclear. Here, we use two-photon calcium imaging in the auditory cortex of behaving mice to test the hypothesis that neural responses to termination of sound sequences ("Off-responses") encode their acoustic history and behavioral salience. We find that auditory cortical Off-responses encode preceding sound sequences and that learning to associate a sound sequence with a reward induces enhancement of Off-responses relative to responses during the sound sequence ("On-responses"). Furthermore, learning enhances network-level discriminability of sound sequences by Off-responses. Last, learning-induced plasticity of Off-responses but not On-responses lasts to the next day. These findings identify auditory cortical Off-responses as a key neural signature of acquired sound-sequence salience.
Collapse
Affiliation(s)
- Joonyeup Lee
- Department of Psychology, University of Michigan, Ann Arbor, MI 48109, USA
| | - Gideon Rothschild
- Department of Psychology, University of Michigan, Ann Arbor, MI 48109, USA; Kresge Hearing Research Institute and Department of Otolaryngology - Head and Neck Surgery, University of Michigan, Ann Arbor, MI 48109, USA.
| |
Collapse
|
19
|
Pouw W, Proksch S, Drijvers L, Gamba M, Holler J, Kello C, Schaefer RS, Wiggins GA. Multilevel rhythms in multimodal communication. Philos Trans R Soc Lond B Biol Sci 2021; 376:20200334. [PMID: 34420378 PMCID: PMC8380971 DOI: 10.1098/rstb.2020.0334] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/05/2021] [Indexed: 12/16/2022] Open
Abstract
It is now widely accepted that the brunt of animal communication is conducted via several modalities, e.g. acoustic and visual, either simultaneously or sequentially. This is a laudable multimodal turn relative to traditional accounts of temporal aspects of animal communication which have focused on a single modality at a time. However, the fields that are currently contributing to the study of multimodal communication are highly varied, and still largely disconnected given their sole focus on a particular level of description or their particular concern with human or non-human animals. Here, we provide an integrative overview of converging findings that show how multimodal processes occurring at neural, bodily, as well as social interactional levels each contribute uniquely to the complex rhythms that characterize communication in human and non-human animals. Though we address findings for each of these levels independently, we conclude that the most important challenge in this field is to identify how processes at these different levels connect. This article is part of the theme issue 'Synchrony and rhythm interaction: from the brain to behavioural ecology'.
Collapse
Affiliation(s)
- Wim Pouw
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Shannon Proksch
- Cognitive and Information Sciences, University of California, Merced, CA, USA
| | - Linda Drijvers
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Marco Gamba
- Department of Life Sciences and Systems Biology, University of Turin, Turin, Italy
| | - Judith Holler
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Christopher Kello
- Cognitive and Information Sciences, University of California, Merced, CA, USA
| | - Rebecca S. Schaefer
- Health, Medical and Neuropsychology unit, Institute for Psychology, Leiden University, Leiden, The Netherlands
- Academy for Creative and Performing Arts, Leiden University, Leiden, The Netherlands
| | - Geraint A. Wiggins
- Vrije Universiteit Brussel, Brussels, Belgium and Queen Mary University of London, UK
- Queen Mary University, London, UK
| |
Collapse
|
20
|
Dotov D, Trainor LJ. Cross-frequency coupling explains the preference for simple ratios in rhythmic behaviour and the relative stability across non-synchronous patterns. Philos Trans R Soc Lond B Biol Sci 2021; 376:20200333. [PMID: 34420377 DOI: 10.1098/rstb.2020.0333] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/18/2022] Open
Abstract
Rhythms are important for understanding coordinated behaviours in ecological systems. The repetitive nature of rhythms affords prediction, planning of movements and coordination of processes within and between individuals. A major challenge is to understand complex forms of coordination when they differ from complete synchronization. By expressing phase as ratio of a cycle, we adapted levels of the Farey tree as a metric of complexity mapped to the range between in-phase and anti-phase synchronization. In a bimanual tapping task, this revealed an increase of variability with ratio complexity, a range of hidden and unstable yet measurable modes, and a rank-frequency scaling law across these modes. We use the phase-attractive circle map to propose an interpretation of these findings in terms of hierarchical cross-frequency coupling (CFC). We also consider the tendency for small-integer attractors in the single-hand repeated tapping of three-interval rhythms reported in the literature. The phase-attractive circle map has wider basins of attractions for such ratios. This work motivates the question whether CFC intrinsic to neural dynamics implements low-level priors for timing and coordination and thus becomes involved in phenomena as diverse as attractor states in bimanual coordination and the cross-cultural tendency for musical rhythms to have simple interval ratios. This article is part of the theme issue 'Synchrony and rhythm interaction: from the brain to behavioural ecology'.
Collapse
Affiliation(s)
- Dobromir Dotov
- LIVELab, McMaster University, 1280 Main Street West, Hamilton, Ontario, Canada L8S4K1.,Psychology, Neuroscience and Behaviour, McMaster University, Ontario, Canada
| | - Laurel J Trainor
- LIVELab, McMaster University, 1280 Main Street West, Hamilton, Ontario, Canada L8S4K1.,Psychology, Neuroscience and Behaviour, McMaster University, Ontario, Canada.,Rotman Research Institute, Toronto, Canada
| |
Collapse
|
21
|
Bloomfield L, Lane E, Mangalam M, Kelty-Stephen DG. Perceiving and remembering speech depend on multifractal nonlinearity in movements producing and exploring speech. J R Soc Interface 2021; 18:20210272. [PMID: 34343455 DOI: 10.1098/rsif.2021.0272] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/04/2023] Open
Abstract
Speech perception and memory for speech require active engagement. Gestural theories have emphasized mainly the effect of speaker's movements on speech perception. They fail to address the effects of listener movement, focusing on communication as a boundary condition constraining movement among interlocutors. The present work attempts to break new ground by using multifractal geometry of physical movement as a common currency for supporting both sides of the speaker-listener dyads. Participants self-paced their listening to a narrative, after which they completed a test of memory querying their narrative comprehension and their ability to recognize words from the story. The multifractal evidence of nonlinear interactions across timescales predicted the fluency of speech perception. Self-pacing movements that enabled listeners to control the presentation of speech sounds constituted a rich exploratory process. The multifractal nonlinearity of this exploration supported several aspects of memory for the perceived spoken language. These findings extend the role of multifractal geometry in the speaker's movements to the narrative case of speech perception. In addition to posing novel basic research questions, these findings make a compelling case for calibrating multifractal structure in text-to-speech synthesizers for better perception and memory of speech.
Collapse
Affiliation(s)
| | - Elizabeth Lane
- Department of Psychology, Grinnell College, Grinnell, IA 50112, USA
| | - Madhur Mangalam
- Department of Physical Therapy, Movement and Rehabilitation Sciences, Northeastern University, Boston, MA 02115, USA
| | | |
Collapse
|
22
|
Abstract
Abstract
The aim of this paper is to review recent hypotheses on the evolutionary origins of music in Homo sapiens, taking into account the most influential traditional hypotheses. To date, theories derived from evolution have focused primarily on the importance that music carries in solving detailed adaptive problems. The three most influential theoretical concepts have described the evolution of human music in terms of 1) sexual selection, 2) the formation of social bonds, or treated it 3) as a byproduct. According to recent proposals, traditional hypotheses are flawed or insufficient in fully explaining the complexity of music in Homo sapiens. This paper will critically discuss three traditional hypotheses of music evolution (music as an effect of sexual selection, a mechanism of social bonding, and a byproduct), as well as and two recent concepts of music evolution - music as a credible signal and Music and Social Bonding (MSB) hypothesis.
Collapse
|
23
|
Kojima H, Chen D, Oka M, Ikegami T. Analysis and Design of Social Presence in a Computer-Mediated Communication System. Front Psychol 2021; 12:641927. [PMID: 34108909 PMCID: PMC8180569 DOI: 10.3389/fpsyg.2021.641927] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2020] [Accepted: 04/19/2021] [Indexed: 12/02/2022] Open
Abstract
Social presence, or the subjective experience of being present with another existing person, varies with the interaction medium. In general, social presence research has mainly focused on uni-directional aspects of each exchanged message, not on bidirectional interactions. Our primary purpose is to introduce such bidirectional evaluation by quantifying the degree of social presence with a few statistical measures. To this end, we developed a software called “TypeTrace” that records all keystrokes of online chat interactants and reenacts their typing actions and analyzed the results from different chat conditions, mainly focusing on the characterization of bi-directional interactions. We also compared the chat interaction patterns with the patterns from phone call datasets to investigate the difference of live communication in different media. The hypothesis of the experiment was that either richness or concurrency of communication is important for organizing social presence. Richness is defined by the variety of information at a time in communication and the concurrency is the number of temporal thread being processed at the same time. Our results show that when we merely increase the richness of information by presenting the typing process, the cognition of others' presence does not significantly increase. However, when the information concurrency is augmented by introducing the transmission of realtime text, we found that the transfer entropy between the interactants becomes considerably higher, and the social presence and emotional arousal, intimacy increased. High transfer entropy was also observed in the phone call dataset. This result shows that the mere augmentation of information richness does not necessarily lead to increased social presence, and concurrent communication is another critical factor for fostering vivid conversation in digital environments.
Collapse
Affiliation(s)
- Hiroki Kojima
- Graduate School of Arts and Sciences, University of Tokyo, Tokyo, Japan
| | - Dominique Chen
- Faculty of Letters and Science, Waseda University, Tokyo, Japan
| | - Mizuki Oka
- Graduate School of Science and Technology, University of Tsukuba, Ibaraki, Japan
| | - Takashi Ikegami
- Graduate School of Arts and Sciences, University of Tokyo, Tokyo, Japan
| |
Collapse
|
24
|
Mohseni M, Gast V, Redies C. Fractality and Variability in Canonical and Non-Canonical English Fiction and in Non-Fictional Texts. Front Psychol 2021; 12:599063. [PMID: 33868078 PMCID: PMC8044424 DOI: 10.3389/fpsyg.2021.599063] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2020] [Accepted: 03/02/2021] [Indexed: 11/13/2022] Open
Abstract
This study investigates global properties of three categories of English text: canonical fiction, non-canonical fiction, and non-fictional texts. The central hypothesis of the study is that there are systematic differences with respect to structural design features between canonical and non-canonical fiction, and between fictional and non-fictional texts. To investigate these differences, we compiled a corpus containing texts of the three categories of interest, the Jena Corpus of Expository and Fictional Prose (JEFP Corpus). Two aspects of global structure are investigated, variability and self-similar (fractal) patterns, which reflect long-range correlations along texts. We use four types of basic observations, (i) the frequency of POS-tags per sentence, (ii) sentence length, (iii) lexical diversity, and (iv) the distribution of topic probabilities in segments of texts. These basic observations are grouped into two more general categories, (a) the lower-level properties (i) and (ii), which are observed at the level of the sentence (reflecting linguistic decoding), and (b) the higher-level properties (iii) and (iv), which are observed at the textual level (reflecting comprehension/integration). The observations for each property are transformed into series, which are analyzed in terms of variance and subjected to Multi-Fractal Detrended Fluctuation Analysis (MFDFA), giving rise to three statistics: (i) the degree of fractality ( H ), (ii) the degree of multifractality ( D ), i.e., the width of the fractal spectrum, and (iii) the degree of asymmetry ( A ) of the fractal spectrum. The statistics thus obtained are compared individually across text categories and jointly fed into a classification model (Support Vector Machine). Our results show that there are in fact differences between the three text categories of interest. In general, lower-level text properties are better discriminators than higher-level text properties. Canonical fictional texts differ from non-canonical ones primarily in terms of variability in lower-level text properties. Fractality seems to be a universal feature of text, slightly more pronounced in non-fictional than in fictional texts. On the basis of our results obtained on the basis of corpus data we point out some avenues for future research leading toward a more comprehensive analysis of textual aesthetics, e.g., using experimental methodologies.
Collapse
Affiliation(s)
- Mahdi Mohseni
- Experimental Aesthetics Group, Institute of Anatomy I, Jena University Hospital, University of Jena, Jena, Germany.,Department of English and American Studies, University of Jena, Jena, Germany
| | - Volker Gast
- Department of English and American Studies, University of Jena, Jena, Germany
| | - Christoph Redies
- Experimental Aesthetics Group, Institute of Anatomy I, Jena University Hospital, University of Jena, Jena, Germany
| |
Collapse
|
25
|
Alviar C, Dale R, Dewitt A, Kello C. Multimodal Coordination of Sound and Movement in Music and Speech. DISCOURSE PROCESSES 2020. [DOI: 10.1080/0163853x.2020.1768500] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
Affiliation(s)
- Camila Alviar
- Cognitive and Information Sciences, University of California, Merced
| | - Rick Dale
- Department of Communication, University of California, Los Angeles
| | - Akeiylah Dewitt
- Cognitive and Information Sciences, University of California, Merced
| | - Christopher Kello
- Cognitive and Information Sciences, University of California, Merced
| |
Collapse
|
26
|
Ravignani A, Dalla Bella S, Falk S, Kello CT, Noriega F, Kotz SA. Rhythm in speech and animal vocalizations: a cross-species perspective. Ann N Y Acad Sci 2019; 1453:79-98. [PMID: 31237365 PMCID: PMC6851814 DOI: 10.1111/nyas.14166] [Citation(s) in RCA: 27] [Impact Index Per Article: 5.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2019] [Revised: 05/14/2019] [Accepted: 05/24/2019] [Indexed: 12/31/2022]
Abstract
Why does human speech have rhythm? As we cannot travel back in time to witness how speech developed its rhythmic properties and why humans have the cognitive skills to process them, we rely on alternative methods to find out. One powerful tool is the comparative approach: studying the presence or absence of cognitive/behavioral traits in other species to determine which traits are shared between species and which are recent human inventions. Vocalizations of many species exhibit temporal structure, but little is known about how these rhythmic structures evolved, are perceived and produced, their biological and developmental bases, and communicative functions. We review the literature on rhythm in speech and animal vocalizations as a first step toward understanding similarities and differences across species. We extend this review to quantitative techniques that are useful for computing rhythmic structure in acoustic sequences and hence facilitate cross-species research. We report links between vocal perception and motor coordination and the differentiation of rhythm based on hierarchical temporal structure. While still far from a complete cross-species perspective of speech rhythm, our review puts some pieces of the puzzle together.
Collapse
Affiliation(s)
- Andrea Ravignani
- Artificial Intelligence LaboratoryVrije Universiteit BrusselBrusselsBelgium
- Institute for Advanced StudyUniversity of AmsterdamAmsterdamthe Netherlands
| | - Simone Dalla Bella
- International Laboratory for BrainMusic and Sound Research (BRAMS)MontréalQuebecCanada
- Department of PsychologyUniversity of MontrealMontréalQuebecCanada
- Department of Cognitive PsychologyWarsawPoland
| | - Simone Falk
- International Laboratory for BrainMusic and Sound Research (BRAMS)MontréalQuebecCanada
- Laboratoire de Phonétique et Phonologie, UMR 7018, CNRS/Université Sorbonne Nouvelle Paris‐3Institut de Linguistique et Phonétique générales et appliquéesParisFrance
| | | | - Florencia Noriega
- Chair for Network DynamicsCenter for Advancing Electronics Dresden (CFAED), TU DresdenDresdenGermany
- CODE University of Applied SciencesBerlinGermany
| | - Sonja A. Kotz
- International Laboratory for BrainMusic and Sound Research (BRAMS)MontréalQuebecCanada
- Basic and Applied NeuroDynamics Laboratory, Faculty of Psychology and Neuroscience, Department of Neuropsychology and PsychopharmacologyMaastricht UniversityMaastrichtthe Netherlands
- Department of NeuropsychologyMax‐Planck Institute for Human Cognitive and Brain SciencesLeipzigGermany
| |
Collapse
|
27
|
Affiliation(s)
- Andrea Ravignani
- Artificial Intelligence Lab, Vrije Universiteit Brussel, Brussels, Belgium
- Research Department, Sealcentre Pieterburen, Pieterburen, The Netherlands
| |
Collapse
|
28
|
Tian S, Fan X, Wang Y, Liu Z, Wang L. An in vitro experimental study on the relationship between pulsatile tinnitus and the dehiscence/thinness of sigmoid sinus cortical plate. J Biomech 2019; 84:197-203. [PMID: 30665710 DOI: 10.1016/j.jbiomech.2018.12.049] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2018] [Revised: 12/25/2018] [Accepted: 12/31/2018] [Indexed: 12/18/2022]
Abstract
Pulsatile tinnitus (PT), characterized as pulse-synchronous, is generally objective. Sigmoid sinus (SS) venous sound is widely suggested to be a possible sound source of PT. The dehiscence and thinness of SS cortical plate (CP) was commonly reported as PT pathology in previous studies, but lack quantitative or biomechanical analysis. In this study, it was aimed to quantify the relationship between venous sound and CP dehiscence/thinness using in vitro experiment. The in vitro models of SS and CP were established based on 3D-printing, with the developed pulsatile venous flow in the SS model. The generated sound signal and the vibration response at the dehiscent/thinned area were analyzed. The sound signal generated in the normal-sized dehiscence model was pulse-synchronous within 100--400 Hz, which had similar acoustic characteristics as the clinical PT sounds. It was concluded that the pulsatile venous sound is produced at TS-SS junction in case of CP dehiscence. The CP, even a thinned one can effectively diminish the venous sound and sound-generating pulsatile vibration at TS-SS junction. The CP dehiscence would induce pulse-synchronous and high pressure venous sound, as well as pulse-synchronous vibration above 20 Hz, regardless of the dehiscence size. On the contrary, the CP thinness would not induce obvious venous sound or pulsatile vibration above 20 Hz.
Collapse
Affiliation(s)
- Shan Tian
- Key Laboratory for Biomechanics and Mechanobiology of Ministry of Education, School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China; Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing 100083, China
| | - Xingyu Fan
- College of Bioengineering, Chongqing University, Chongqing 400044, China
| | - Yawei Wang
- Key Laboratory for Biomechanics and Mechanobiology of Ministry of Education, School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China; Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing 100083, China
| | - Zhaohui Liu
- Department of Radiology, Beijing Tongren Hospital, Capital Medical University, Beijing 100730, China.
| | - Lizhen Wang
- Key Laboratory for Biomechanics and Mechanobiology of Ministry of Education, School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China; Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing 100083, China.
| |
Collapse
|
29
|
Kotz S, Ravignani A, Fitch W. The Evolution of Rhythm Processing. Trends Cogn Sci 2018; 22:896-910. [DOI: 10.1016/j.tics.2018.08.002] [Citation(s) in RCA: 68] [Impact Index Per Article: 11.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2018] [Revised: 07/25/2018] [Accepted: 08/02/2018] [Indexed: 01/14/2023]
|
30
|
Ward RM, Kelty-Stephen DG. Bringing the Nonlinearity of the Movement System to Gestural Theories of Language Use: Multifractal Structure of Spoken English Supports the Compensation for Coarticulation in Human Speech Perception. Front Physiol 2018; 9:1152. [PMID: 30233386 PMCID: PMC6129613 DOI: 10.3389/fphys.2018.01152] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2018] [Accepted: 07/31/2018] [Indexed: 01/13/2023] Open
Abstract
Coarticulation is the tendency for speech vocalization and articulation even at the phonemic level to change with context, and compensation for coarticulation (CfC) reflects the striking human ability to perceive phonemic stability despite this variability. A current controversy centers on whether CfC depends on contrast between formants of a speech-signal spectrogram-specifically, contrast between offset formants concluding context stimuli and onset formants opening the target sound-or on speech-sound variability specific to the coordinative movement of speech articulators (e.g., vocal folds, postural muscles, lips, tongues). This manuscript aims to encode that coordinative-movement context in terms of speech-signal multifractal structure and to determine whether speech's multifractal structure might explain the crucial gestural support for any proposed spectral contrast. We asked human participants to categorize individual target stimuli drawn from an 11-step [ga]-to-[da] continuum as either phonemes "GA" or "DA." Three groups each heard a specific-type context stimulus preceding target stimuli: either real-speech [al] or [a], sine-wave tones at the third-formant offset frequency of either [al] or [aɹ], and either simulated-speech contexts [al] or [aɹ]. Here, simulating speech contexts involved randomizing the sequence of relatively homogeneous pitch periods within vowel-sound [a] of each [al] and [aɹ]. Crucially, simulated-speech contexts had the same offset and extremely similar vowel formants as and, to additional naïve participants, sounded identical to real-speech contexts. However, randomization distorted original speech-context multifractality, and effects of spectral contrast following speech only appeared after regression modeling of trial-by-trial "GA" judgments controlled for context-stimulus multifractality. Furthermore, simulated-speech contexts elicited faster responses (like tone contexts do) and weakened known biases in CfC, suggesting that spectral contrast depends on the nonlinear interactions across multiple scales that articulatory gestures express through the speech signal. Traditional mouse-tracking behaviors measured as participants moved their computer-mouse cursor to register their "GA"-or-"DA" decisions with mouse-clicks suggest that listening to speech leads the movement system to resonate with the multifractality of context stimuli. We interpret these results as shedding light on a new multifractal terrain upon which to found a better understanding in which movement systems play an important role in shaping how speech perception makes use of acoustic information.
Collapse
|
31
|
Ravignani A, Kello CT, de Reus K, Kotz SA, Dalla Bella S, Méndez-Aróstegui M, Rapado-Tamarit B, Rubio-Garcia A, de Boer B. Ontogeny of vocal rhythms in harbor seal pups: an exploratory study. Curr Zool 2018; 65:107-120. [PMID: 30697246 PMCID: PMC6347067 DOI: 10.1093/cz/zoy055] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2018] [Accepted: 07/02/2018] [Indexed: 11/15/2022] Open
Abstract
Puppyhood is a very active social and vocal period in a harbor seal’s life Phoca vitulina. An important feature of vocalizations is their temporal and rhythmic structure, and understanding vocal timing and rhythms in harbor seals is critical to a cross-species hypothesis in evolutionary neuroscience that links vocal learning, rhythm perception, and synchronization. This study utilized analytical techniques that may best capture rhythmic structure in pup vocalizations with the goal of examining whether (1) harbor seal pups show rhythmic structure in their calls and (2) rhythms evolve over time. Calls of 3 wild-born seal pups were recorded daily over the course of 1–3 weeks; 3 temporal features were analyzed using 3 complementary techniques. We identified temporal and rhythmic structure in pup calls across different time windows. The calls of harbor seal pups exhibit some degree of temporal and rhythmic organization, which evolves over puppyhood and resembles that of other species’ interactive communication. We suggest next steps for investigating call structure in harbor seal pups and propose comparative hypotheses to test in other pinniped species.
Collapse
Affiliation(s)
- Andrea Ravignani
- Research Department, Sealcentre Pieterburen, Pieterburen, The Netherlands.,Artificial Intelligence Lab, Vrije Universiteit Brussel, Brussels, Belgium
| | - Christopher T Kello
- Cognitive and Information Sciences, University of California, Merced, CA, USA
| | - Koen de Reus
- Research Department, Sealcentre Pieterburen, Pieterburen, The Netherlands
| | - Sonja A Kotz
- Basic and Applied NeuroDynamics Lab, Faculty of Psychology and Neuroscience, Department of Neuropsychology & Psychopharmacology, Maastricht University, Maastricht, The Netherlands.,Department of Neuropsychology, Max-Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.,International Laboratory for Brain, Music and Sound Research (BRAMS), Montréal, QC, Canada
| | - Simone Dalla Bella
- International Laboratory for Brain, Music and Sound Research (BRAMS), Montréal, QC, Canada.,Department of Psychology, University of Montreal, Montréal, QC, Canada.,Department of Cognitive Psychology, WSFiZ in Warsaw, Warsaw, Poland
| | | | | | - Ana Rubio-Garcia
- Research Department, Sealcentre Pieterburen, Pieterburen, The Netherlands
| | - Bart de Boer
- Artificial Intelligence Lab, Vrije Universiteit Brussel, Brussels, Belgium
| |
Collapse
|
32
|
Abney DH, Dale R, Louwerse MM, Kello CT. The Bursts and Lulls of Multimodal Interaction: Temporal Distributions of Behavior Reveal Differences Between Verbal and Non-Verbal Communication. Cogn Sci 2018; 42:1297-1316. [PMID: 29630740 DOI: 10.1111/cogs.12612] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2017] [Revised: 11/22/2017] [Accepted: 01/26/2018] [Indexed: 11/30/2022]
Abstract
Recent studies of naturalistic face-to-face communication have demonstrated coordination patterns such as the temporal matching of verbal and non-verbal behavior, which provides evidence for the proposal that verbal and non-verbal communicative control derives from one system. In this study, we argue that the observed relationship between verbal and non-verbal behaviors depends on the level of analysis. In a reanalysis of a corpus of naturalistic multimodal communication (Louwerse, Dale, Bard, & Jeuniaux, ), we focus on measuring the temporal patterns of specific communicative behaviors in terms of their burstiness. We examined burstiness estimates across different roles of the speaker and different communicative modalities. We observed more burstiness for verbal versus non-verbal channels, and for more versus less informative language subchannels. Using this new method for analyzing temporal patterns in communicative behaviors, we show that there is a complex relationship between verbal and non-verbal channels. We propose a "temporal heterogeneity" hypothesis to explain how the language system adapts to the demands of dialog.
Collapse
Affiliation(s)
- Drew H Abney
- Department of Psychological and Brain Sciences, Indiana University
| | - Rick Dale
- Department of Communication, University of California, Los Angeles
| | - Max M Louwerse
- Cognitive Science and Artificial Intelligence, Tilburg University
| | | |
Collapse
|
33
|
Schneider JN, Mercado E. Characterizing the rhythm and tempo of sound production by singing whales. BIOACOUSTICS 2018. [DOI: 10.1080/09524622.2018.1428827] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
Affiliation(s)
| | - Eduardo Mercado
- Department of Psychology, University at Buffalo, The State University of New York, Buffalo, NY, USA
| |
Collapse
|
34
|
Ravignani A. Spontaneous rhythms in a harbor seal pup calls. BMC Res Notes 2018; 11:3. [PMID: 29298731 PMCID: PMC5751680 DOI: 10.1186/s13104-017-3107-6] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2017] [Accepted: 12/19/2017] [Indexed: 12/25/2022] Open
Abstract
Objectives Timing and rhythm (i.e. temporal structure) are crucial, though historically neglected, dimensions of animal communication. When investigating these in non-human animals, it is often difficult to balance experimental control and ecological validity. Here I present the first step of an attempt to balance the two, focusing on the timing of vocal rhythms in a harbor seal pup (Phoca vitulina). Collection of this data had a clear aim: To find spontaneous vocal rhythms in this individual in order to design individually-adapted and ecologically-relevant stimuli for a later playback experiment. Data description The calls of one seal pup were recorded. The audio recordings were annotated using Praat, a free software to analyze vocalizations in humans and other animals. The annotated onsets and offsets of vocalizations were then imported in a Python script. The script extracted three types of timing information: the duration of calls, the intervals between calls’ onsets, and the intervals between calls’ maximum-intensity peaks. Based on the annotated data, available to download, I provide simple descriptive statistics for these temporal measures, and compare their distributions.
Collapse
Affiliation(s)
- Andrea Ravignani
- Research Department, Sealcentre Pieterburen, Hoofdstraat 94a, 9968 AG, Pieterburen, The Netherlands. .,Artificial Intelligence Lab, Vrije Universiteit Brussel, Pleinlaan 2, 1050, Brussels, Belgium. .,Language and Cognition Department, Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525 XD, Nijmegen, The Netherlands.
| |
Collapse
|