Behavior Research Methods,
Journal Year:
2025,
Volume and Issue:
57(5)
Published: April 11, 2025
Abstract
The
human
voice
is
remarkably
versatile
and
can
vary
greatly
in
sound
depending
on
how
it
used.
An
increasing
number
of
studies
have
addressed
the
differences
similarities
between
singing
speaking
voice.
However,
finding
adequate
stimuli
material
that
at
same
time
controlled
ecologically
valid
challenging,
most
datasets
lack
variability
terms
vocal
styles
performed
by
Here,
we
describe
a
curated
stimulus
set
vocalizations
where
22
female
singers
melody
excerpts
three
contrasting
(as
lullaby,
as
pop
song,
an
opera
aria)
spoke
text
aloud
two
if
to
adult
or
infant).
All
productions
were
made
with
songs’
original
lyrics,
Brazilian
Portuguese,
a/lu/sound.
This
dataset
1320
was
validated
through
forced-choice
lab
experiment
(
N
=
25
for
each
stimulus)
lay
listeners
could
recognize
intended
vocalization
style
high
accuracy
(proportion
correct
recognition
superior
69%
all
styles).
We
also
provide
acoustic
characterization
stimuli,
depicting
clear
profiles
vocalization.
recordings
are
freely
available
under
Creative
Commons
license
be
downloaded
https://osf.io/cgexn/
.
Scientific Reports,
Journal Year:
2024,
Volume and Issue:
14(1)
Published: Jan. 8, 2024
Music
and
speech
are
encountered
daily
unique
to
human
beings.
Both
transformed
by
the
auditory
pathway
from
an
initial
acoustical
encoding
higher
level
cognition.
Studies
of
cortex
have
revealed
distinct
brain
responses
music
speech,
but
differences
may
emerge
in
or
be
inherited
different
subcortical
encoding.
In
first
part
this
study,
we
derived
brainstem
response
(ABR),
a
measure
encoding,
recorded
using
two
analysis
methods.
The
method,
described
previously
acoustically
based,
yielded
very
ABRs
between
sound
classes.
second
however,
developed
here
based
on
physiological
model
periphery,
gave
highly
correlated
speech.
We
determined
superiority
method
through
several
metrics,
suggesting
there
is
no
appreciable
impact
stimulus
class
(i.e.,
vs
speech)
way
acoustics
encoded
subcortically.
study's
part,
considered
cortex.
Our
new
resulted
cortical
becoming
more
similar
with
remaining
differences.
results
taken
together
suggest
that
evidence
for
stimulus-class
dependent
processing
at
not
level.
Frontiers in Cognition,
Journal Year:
2025,
Volume and Issue:
4
Published: Feb. 24, 2025
This
article
addresses
the
origins
and
purpose
of
communal
music-making,
including
dance,
its
role
in
human
sociality.
It
accords
special
significance
to
adapted
nature
vocalization,
sensorimotor
discrimination
that
allows
prediction
then
generation
musically
relevant,
coordinated
simultaneous
movements.
Commencing
with
a
historical
survey
development
ideas
about
evolutionary
importance
music
social
behavior,
this
mini-review
sets
out
define
explore
key
issues
involved
an
explanation.
These
include:
acquisition
control
parameters
required
for
vocal
production
(synchronization
pitch,
timbre,
duration
loudness);
exchange
transmission
pitched
utterances
unison
as
well
harmony;
roles
natural
sexual
selection
shaping
musical
abilities;
cooperative
consequences
bonding
such
interaction
throughout
life;
behaviors
across
generations,
between
genes
culture
drives
evolution
complex
behavior
Homo
sapiens
.
The
concludes
brief
review
current
research
deals
contributory
features
field,
especially
neuroscience
which
continues
provide
important
psychophysiological
data
reinforces
long-held
proposal
has
promoting
cooperative,
prosocial
interactions
leading
health
wellbeing
over
lifespan.
Neurobiology of Language,
Journal Year:
2022,
Volume and Issue:
3(4), P. 615 - 664
Published: Jan. 1, 2022
Abstract
Using
individual
differences
approaches,
a
growing
body
of
literature
finds
positive
associations
between
musicality
and
language-related
abilities,
complementing
prior
findings
links
musical
training
language
skills.
Despite
these
associations,
has
been
often
overlooked
in
mainstream
models
acquisition
development.
To
better
understand
the
biological
basis
differences,
we
propose
Musical
Abilities,
Pleiotropy,
Language,
Environment
(MAPLE)
framework.
This
novel
integrative
framework
posits
that
abilities
likely
share
some
common
genetic
architecture
(i.e.,
pleiotropy)
addition
to
degree
overlapping
neural
endophenotypes,
influences
on
musically
linguistically
enriched
environments.
Drawing
upon
recent
advances
genomic
methodologies
for
unraveling
pleiotropy,
outline
testable
predictions
future
research
development
how
its
underlying
neurobiological
substrates
may
be
supported
by
pleiotropy
with
musicality.
In
support
MAPLE
framework,
review
discuss
from
over
seventy
behavioral
studies,
highlighting
is
robustly
associated
range
speech-language
skills
required
communication
These
include
speech
perception-in-noise,
prosodic
perception,
morphosyntactic
skills,
phonological
reading
aspects
second/foreign
learning.
Overall,
current
work
provides
clear
agenda
studying
musicality-language
using
an
emphasis
leveraging
genomics
complex
traits.
bioRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2023,
Volume and Issue:
unknown
Published: Jan. 29, 2023
Abstract
Humans
produce
two
forms
of
cognitively
complex
vocalizations:
speech
and
song.
It
is
debated
whether
these
differ
based
primarily
on
culturally
specific,
learned
features,
or
if
acoustical
features
can
reliably
distinguish
them.
We
study
the
spectro-temporal
modulation
patterns
vocalizations
produced
by
369
people
living
in
21
urban,
rural,
small-scale
societies
across
six
continents.
Specific
ranges
spectral
temporal
modulations,
overlapping
within
categories
societies,
significantly
differentiate
from
Machine-learning
classification
shows
that
this
effect
cross-culturally
robust,
being
classified
solely
their
all
societies.
Listeners
unfamiliar
with
cultures
classify
using
similar
cues
as
machine
learning
algorithm.
Finally,
are
better
able
to
discriminate
song
than
a
broad
range
other
variables,
suggesting
modulation—a
key
feature
auditory
neuronal
tuning—accounts
for
fundamental
difference
between
categories.
Two-Sentence
Summary
What
distinguishes
singing
speaking?
The
authors
show
consistent
sufficient
different
throughout
world.
PLoS Biology,
Journal Year:
2024,
Volume and Issue:
22(5), P. e3002631 - e3002631
Published: May 28, 2024
Music
and
speech
are
complex
distinct
auditory
signals
that
both
foundational
to
the
human
experience.
The
mechanisms
underpinning
each
domain
widely
investigated.
However,
what
perceptual
mechanism
transforms
a
sound
into
music
or
how
basic
acoustic
information
is
required
distinguish
between
them
remain
open
questions.
Here,
we
hypothesized
sound’s
amplitude
modulation
(AM),
an
essential
temporal
feature
driving
system
across
processing
levels,
critical
for
distinguishing
speech.
Specifically,
in
contrast
paradigms
using
naturalistic
(that
can
be
challenging
interpret),
used
noise-probing
approach
untangle
mechanism:
If
AM
rate
regularity
perceptually
speech,
judging
artificially
noise-synthesized
ambiguous
audio
should
align
with
their
parameters.
Across
4
experiments
(
N
=
335),
higher
peak
frequency
tend
judged
as
lower
music.
Interestingly,
this
principle
consistently
by
all
listeners
judgments,
but
only
musically
sophisticated
In
addition,
more
regular
over
judgment,
regardless
of
musical
sophistication.
data
suggest
rely
on
low-level
property
from
simple
provokes
neurophysiological
evolutionary
speculations.
bioRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2021,
Volume and Issue:
unknown
Published: June 1, 2021
Abstract
Language
and
music
are
two
human-unique
capacities
whose
relationship
remains
debated.
Some
have
argued
for
overlap
in
processing
mechanisms,
especially
structure
processing.
Such
claims
often
concern
the
inferior
frontal
component
of
language
system
located
within
‘Broca’s
area’.
However,
others
failed
to
find
overlap.
Using
a
robust
individual-subject
fMRI
approach,
we
examined
responses
brain
regions
stimuli,
probed
musical
abilities
individuals
with
severe
aphasia.
Across
four
experiments,
obtained
clear
answer:
perception
does
not
engage
system,
judgments
about
possible
even
presence
damage
network.
In
particular,
regions’
generally
low,
below
fixation
baseline,
never
exceed
elicited
by
non-music
auditory
conditions,
like
animal
sounds.
Further,
sensitive
structure:
they
show
low
intact
structure-scrambled
music,
melodies
vs.
without
structural
violations.
Finally,
line
past
patient
investigations,
aphasia
who
cannot
judge
sentence
grammaticality
perform
well
on
melody
well-formedness
judgments.
Thus
mechanisms
that
process
do
appear
including
syntax.
PLoS Biology,
Journal Year:
2022,
Volume and Issue:
20(7), P. e3001675 - e3001675
Published: July 28, 2022
The
ability
to
recognize
abstract
features
of
voice
during
auditory
perception
is
an
intricate
feat
human
audition.
For
the
listener,
this
occurs
in
near-automatic
fashion
seamlessly
extract
complex
cues
from
a
highly
variable
signal.
Voice
depends
on
specialized
regions
cortex,
including
superior
temporal
gyrus
(STG)
and
sulcus
(STS).
However,
nature
encoding
at
cortical
level
remains
poorly
understood.
We
leverage
intracerebral
recordings
across
cortex
presentation
nonvoice
acoustic
stimuli
examine
8
patient-participants
undergoing
epilepsy
surgery
evaluation.
show
that
selectivity
increases
along
hierarchy
supratemporal
plane
(STP)
STG
STS.
Results
accurate
decoding
vocalizations
activity
even
complete
absence
linguistic
content.
These
findings
early,
less-selective
window
neural
STS
followed
by
sustained,
strongly
voice-selective
window.
Encoding
models
demonstrate
divergence
hierarchy,
wherein
STG/STS
responses
are
best
explained
category
acoustics,
as
opposed
alone.
This
contrast
recorded
STP,
which
were
accounted
for
features.
support
model
engages
categorical
mechanisms
within
facilitate
feature
extraction.
iScience,
Journal Year:
2023,
Volume and Issue:
26(7), P. 107223 - 107223
Published: June 28, 2023
Language
and
music
involve
the
productive
combination
of
basic
units
into
structures.
It
remains
unclear
whether
brain
regions
sensitive
to
linguistic
musical
structure
are
co-localized.
We
report
an
intraoperative
awake
craniotomy
in
which
a
left-hemispheric
language-dominant
professional
musician
underwent
cortical
stimulation
mapping
(CSM)
electrocorticography
language
perception
production
during
repetition
tasks.
Musical
sequences
were
melodic
or
amelodic,
differed
algorithmic
compressibility
(Lempel-Ziv
complexity).
Auditory
recordings
sentences
syntactic
complexity
(single
vs.
multiple
phrasal
embeddings).
CSM
posterior
superior
temporal
gyrus
(pSTG)
disrupted
production,
along
with
speech
production.
pSTG
middle
(pMTG)
activated
for
(broadband
gamma;
70-150
Hz).
pMTG
activity
was
modulated
by
complexity,
while
complexity.
This
points
shared
resources
comprehension,
but
distinct
neural
signatures
processing
domain-specific
structural
features.
To
what
extent
does
speech
and
music
processing
rely
on
domain-specific
domain-general
neural
networks?
Using
whole-brain
intracranial
EEG
recordings
in
18
epilepsy
patients
listening
to
natural,
continuous
or
music,
we
investigated
the
presence
of
frequency-specific
network-level
brain
activity.
We
combined
it
with
a
statistical
approach
which
clear
operational
distinction
is
made
between
shared,
preferred,
domain-selective
responses.
show
that
majority
focal
activity
shared
processing.
Our
data
also
reveal
an
absence
anatomical
regional
selectivity.
Instead,
responses
are
restricted
distributed
coherent
oscillations,
typical
spectral
fingerprints.
work
highlights
importance
considering
natural
stimuli
dynamics
their
full
complexity
map
cognitive
functions.
European Journal of Neuroscience,
Journal Year:
2023,
Volume and Issue:
59(9), P. 2193 - 2207
Published: April 29, 2023
Abstract
Pupil
size
covaries
with
the
diffusion
rate
of
cholinergic
and
noradrenergic
neurons
throughout
brain,
which
are
essential
to
arousal.
Recent
findings
suggest
that
slow
pupil
fluctuations
during
locomotion
an
index
sustained
activity
in
axons,
whereas
phasic
dilations
related
axons.
Here,
we
investigated
movement
induced
arousal
(i.e.,
by
singing
swaying
music),
hypothesising
actively
engaging
musical
behaviour
will
provoke
stronger
emotional
engagement
participants
lead
different
qualitative
patterns
tonic
activity.
A
challenge
analysis
data
is
turbulent
diameter
due
exogenous
ocular
commonly
encountered
motor
tasks
high
variability
typically
found
between
individuals.
To
address
this,
developed
algorithm
adaptively
estimates
removes
responses
events,
as
well
a
functional
methodology,
derived
from
Pfaffs'
generalised
arousal,
provides
new
statistical
dimension
on
how
can
be
interpreted
according
putative
neuromodulatory
signalling.
We
enhanced
cholinergic‐related
having
opportunity
move
your
body
while
performing
amplified
effect
Phasic
oscillations
execution
attenuated
time,
often
measure
sense
agency
over
movement.