CoVox: A dataset of contrasting vocalizations DOI Creative Commons
Camila Bruder, Pauline Larrouy-Maestri

Behavior Research Methods, Journal Year: 2025, Volume and Issue: 57(5)

Published: April 11, 2025

Abstract The human voice is remarkably versatile and can vary greatly in sound depending on how it used. An increasing number of studies have addressed the differences similarities between singing speaking voice. However, finding adequate stimuli material that at same time controlled ecologically valid challenging, most datasets lack variability terms vocal styles performed by Here, we describe a curated stimulus set vocalizations where 22 female singers melody excerpts three contrasting (as lullaby, as pop song, an opera aria) spoke text aloud two if to adult or infant). All productions were made with songs’ original lyrics, Brazilian Portuguese, a/lu/sound. This dataset 1320 was validated through forced-choice lab experiment ( N = 25 for each stimulus) lay listeners could recognize intended vocalization style high accuracy (proportion correct recognition superior 69% all styles). We also provide acoustic characterization stimuli, depicting clear profiles vocalization. recordings are freely available under Creative Commons license be downloaded https://osf.io/cgexn/ .

Language: Английский

Subcortical responses to music and speech are alike while cortical responses diverge DOI Creative Commons
Shan Tong, Madeline S. Cappelloni, Ross K. Maddox

et al.

Scientific Reports, Journal Year: 2024, Volume and Issue: 14(1)

Published: Jan. 8, 2024

Music and speech are encountered daily unique to human beings. Both transformed by the auditory pathway from an initial acoustical encoding higher level cognition. Studies of cortex have revealed distinct brain responses music speech, but differences may emerge in or be inherited different subcortical encoding. In first part this study, we derived brainstem response (ABR), a measure encoding, recorded using two analysis methods. The method, described previously acoustically based, yielded very ABRs between sound classes. second however, developed here based on physiological model periphery, gave highly correlated speech. We determined superiority method through several metrics, suggesting there is no appreciable impact stimulus class (i.e., vs speech) way acoustics encoded subcortically. study's part, considered cortex. Our new resulted cortical becoming more similar with remaining differences. results taken together suggest that evidence for stimulus-class dependent processing at not level.

Language: Английский

Citations

12

Music as a social instrument: a brief historical and conceptual perspective DOI Creative Commons
Nicholas Bannan,

Alan R. Harvey

Frontiers in Cognition, Journal Year: 2025, Volume and Issue: 4

Published: Feb. 24, 2025

This article addresses the origins and purpose of communal music-making, including dance, its role in human sociality. It accords special significance to adapted nature vocalization, sensorimotor discrimination that allows prediction then generation musically relevant, coordinated simultaneous movements. Commencing with a historical survey development ideas about evolutionary importance music social behavior, this mini-review sets out define explore key issues involved an explanation. These include: acquisition control parameters required for vocal production (synchronization pitch, timbre, duration loudness); exchange transmission pitched utterances unison as well harmony; roles natural sexual selection shaping musical abilities; cooperative consequences bonding such interaction throughout life; behaviors across generations, between genes culture drives evolution complex behavior Homo sapiens . The concludes brief review current research deals contributory features field, especially neuroscience which continues provide important psychophysiological data reinforces long-held proposal has promoting cooperative, prosocial interactions leading health wellbeing over lifespan.

Language: Английский

Citations

1

The Musical Abilities, Pleiotropy, Language, and Environment (MAPLE) Framework for Understanding Musicality-Language Links Across the Lifespan DOI Creative Commons
Srishti Nayak, Peyton L. Coleman, Enikő Ladányi

et al.

Neurobiology of Language, Journal Year: 2022, Volume and Issue: 3(4), P. 615 - 664

Published: Jan. 1, 2022

Abstract Using individual differences approaches, a growing body of literature finds positive associations between musicality and language-related abilities, complementing prior findings links musical training language skills. Despite these associations, has been often overlooked in mainstream models acquisition development. To better understand the biological basis differences, we propose Musical Abilities, Pleiotropy, Language, Environment (MAPLE) framework. This novel integrative framework posits that abilities likely share some common genetic architecture (i.e., pleiotropy) addition to degree overlapping neural endophenotypes, influences on musically linguistically enriched environments. Drawing upon recent advances genomic methodologies for unraveling pleiotropy, outline testable predictions future research development how its underlying neurobiological substrates may be supported by pleiotropy with musicality. In support MAPLE framework, review discuss from over seventy behavioral studies, highlighting is robustly associated range speech-language skills required communication These include speech perception-in-noise, prosodic perception, morphosyntactic skills, phonological reading aspects second/foreign learning. Overall, current work provides clear agenda studying musicality-language using an emphasis leveraging genomics complex traits.

Language: Английский

Citations

38

Spectro-temporal acoustical markers differentiate speech from song across cultures DOI Open Access
Philippe Albouy, Samuel A. Mehr, Roxane S. Hoyer

et al.

bioRxiv (Cold Spring Harbor Laboratory), Journal Year: 2023, Volume and Issue: unknown

Published: Jan. 29, 2023

Abstract Humans produce two forms of cognitively complex vocalizations: speech and song. It is debated whether these differ based primarily on culturally specific, learned features, or if acoustical features can reliably distinguish them. We study the spectro-temporal modulation patterns vocalizations produced by 369 people living in 21 urban, rural, small-scale societies across six continents. Specific ranges spectral temporal modulations, overlapping within categories societies, significantly differentiate from Machine-learning classification shows that this effect cross-culturally robust, being classified solely their all societies. Listeners unfamiliar with cultures classify using similar cues as machine learning algorithm. Finally, are better able to discriminate song than a broad range other variables, suggesting modulation—a key feature auditory neuronal tuning—accounts for fundamental difference between categories. Two-Sentence Summary What distinguishes singing speaking? The authors show consistent sufficient different throughout world.

Language: Английский

Citations

19

The human auditory system uses amplitude modulation to distinguish music from speech DOI Creative Commons
Andrew Chang, Xiangbin Teng, M. Florencia Assaneo

et al.

PLoS Biology, Journal Year: 2024, Volume and Issue: 22(5), P. e3002631 - e3002631

Published: May 28, 2024

Music and speech are complex distinct auditory signals that both foundational to the human experience. The mechanisms underpinning each domain widely investigated. However, what perceptual mechanism transforms a sound into music or how basic acoustic information is required distinguish between them remain open questions. Here, we hypothesized sound’s amplitude modulation (AM), an essential temporal feature driving system across processing levels, critical for distinguishing speech. Specifically, in contrast paradigms using naturalistic (that can be challenging interpret), used noise-probing approach untangle mechanism: If AM rate regularity perceptually speech, judging artificially noise-synthesized ambiguous audio should align with their parameters. Across 4 experiments ( N = 335), higher peak frequency tend judged as lower music. Interestingly, this principle consistently by all listeners judgments, but only musically sophisticated In addition, more regular over judgment, regardless of musical sophistication. data suggest rely on low-level property from simple provokes neurophysiological evolutionary speculations.

Language: Английский

Citations

7

The human language system, including its inferior frontal component in ‘Broca’s area’, does not support music perception DOI Creative Commons
Xuanyi Chen, Josef Affourtit, Rachel Ryskin

et al.

bioRxiv (Cold Spring Harbor Laboratory), Journal Year: 2021, Volume and Issue: unknown

Published: June 1, 2021

Abstract Language and music are two human-unique capacities whose relationship remains debated. Some have argued for overlap in processing mechanisms, especially structure processing. Such claims often concern the inferior frontal component of language system located within ‘Broca’s area’. However, others failed to find overlap. Using a robust individual-subject fMRI approach, we examined responses brain regions stimuli, probed musical abilities individuals with severe aphasia. Across four experiments, obtained clear answer: perception does not engage system, judgments about possible even presence damage network. In particular, regions’ generally low, below fixation baseline, never exceed elicited by non-music auditory conditions, like animal sounds. Further, sensitive structure: they show low intact structure-scrambled music, melodies vs. without structural violations. Finally, line past patient investigations, aphasia who cannot judge sentence grammaticality perform well on melody well-formedness judgments. Thus mechanisms that process do appear including syntax.

Language: Английский

Citations

35

Neural responses in human superior temporal cortex support coding of voice representations DOI Creative Commons
Kyle Rupp, Jasmine L. Hect, Madison Remick

et al.

PLoS Biology, Journal Year: 2022, Volume and Issue: 20(7), P. e3001675 - e3001675

Published: July 28, 2022

The ability to recognize abstract features of voice during auditory perception is an intricate feat human audition. For the listener, this occurs in near-automatic fashion seamlessly extract complex cues from a highly variable signal. Voice depends on specialized regions cortex, including superior temporal gyrus (STG) and sulcus (STS). However, nature encoding at cortical level remains poorly understood. We leverage intracerebral recordings across cortex presentation nonvoice acoustic stimuli examine 8 patient-participants undergoing epilepsy surgery evaluation. show that selectivity increases along hierarchy supratemporal plane (STP) STG STS. Results accurate decoding vocalizations activity even complete absence linguistic content. These findings early, less-selective window neural STS followed by sustained, strongly voice-selective window. Encoding models demonstrate divergence hierarchy, wherein STG/STS responses are best explained category acoustics, as opposed alone. This contrast recorded STP, which were accounted for features. support model engages categorical mechanisms within facilitate feature extraction.

Language: Английский

Citations

23

Intraoperative cortical localization of music and language reveals signatures of structural complexity in posterior temporal cortex DOI Creative Commons
Meredith J. McCarty, Elliot Murphy, Xavier Scherschligt

et al.

iScience, Journal Year: 2023, Volume and Issue: 26(7), P. 107223 - 107223

Published: June 28, 2023

Language and music involve the productive combination of basic units into structures. It remains unclear whether brain regions sensitive to linguistic musical structure are co-localized. We report an intraoperative awake craniotomy in which a left-hemispheric language-dominant professional musician underwent cortical stimulation mapping (CSM) electrocorticography language perception production during repetition tasks. Musical sequences were melodic or amelodic, differed algorithmic compressibility (Lempel-Ziv complexity). Auditory recordings sentences syntactic complexity (single vs. multiple phrasal embeddings). CSM posterior superior temporal gyrus (pSTG) disrupted production, along with speech production. pSTG middle (pMTG) activated for (broadband gamma; 70-150 Hz). pMTG activity was modulated by complexity, while complexity. This points shared resources comprehension, but distinct neural signatures processing domain-specific structural features.

Language: Английский

Citations

13

Speech and music recruit frequency-specific distributed and overlapping cortical networks DOI Creative Commons
Noémie te Rietmolen, Manuel Mercier, Agnès Trébuchon

et al.

eLife, Journal Year: 2024, Volume and Issue: 13

Published: Feb. 8, 2024

To what extent does speech and music processing rely on domain-specific domain-general neural networks? Using whole-brain intracranial EEG recordings in 18 epilepsy patients listening to natural, continuous or music, we investigated the presence of frequency-specific network-level brain activity. We combined it with a statistical approach which clear operational distinction is made between shared, preferred, domain-selective responses. show that majority focal activity shared processing. Our data also reveal an absence anatomical regional selectivity. Instead, responses are restricted distributed coherent oscillations, typical spectral fingerprints. work highlights importance considering natural stimuli dynamics their full complexity map cognitive functions.

Language: Английский

Citations

5

Cholinergic‐related pupil activity reflects level of emotionality during motor performance DOI Creative Commons
Marc Vidal, Kelsey E. Onderdijk, Ana M. Aguilera

et al.

European Journal of Neuroscience, Journal Year: 2023, Volume and Issue: 59(9), P. 2193 - 2207

Published: April 29, 2023

Abstract Pupil size covaries with the diffusion rate of cholinergic and noradrenergic neurons throughout brain, which are essential to arousal. Recent findings suggest that slow pupil fluctuations during locomotion an index sustained activity in axons, whereas phasic dilations related axons. Here, we investigated movement induced arousal (i.e., by singing swaying music), hypothesising actively engaging musical behaviour will provoke stronger emotional engagement participants lead different qualitative patterns tonic activity. A challenge analysis data is turbulent diameter due exogenous ocular commonly encountered motor tasks high variability typically found between individuals. To address this, developed algorithm adaptively estimates removes responses events, as well a functional methodology, derived from Pfaffs' generalised arousal, provides new statistical dimension on how can be interpreted according putative neuromodulatory signalling. We enhanced cholinergic‐related having opportunity move your body while performing amplified effect Phasic oscillations execution attenuated time, often measure sense agency over movement.

Language: Английский

Citations

11