bioRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2022,
Volume and Issue:
unknown
Published: Aug. 15, 2022
Abstract
The
human
brain
tracks
the
temporal
envelope
of
speech,
which
contains
essential
cues
for
speech
understanding.
Linear
models
are
most
common
tool
to
study
neural
tracking.
However,
information
on
how
is
processed
can
be
lost
since
nonlinear
relations
precluded.
As
an
alternative,
mutual
(MI)
analysis
detect
both
linear
and
relations.
Yet,
several
different
approaches
calculating
MI
applied
without
consensus
approach
use.
Furthermore,
added
value
techniques
remains
a
subject
debate
in
field.
To
resolve
this,
we
analyses
electroencephalography
(EEG)
data
participants
listening
continuous
speech.
Comparing
approaches,
conclude
that
results
reliable
robust
using
Gaussian
copula
approach,
first
transforms
standard
Gaussians.
With
this
valid
technique
studying
Like
models,
it
allows
spatial
interpretations
processing,
peak
latency
analyses,
applications
multiple
EEG
channels
combined.
Finally,
demonstrate
components
single-subject
level,
beyond
limits
models.
We
more
informative
Significance
statement
In
present
study,
addressed
key
methodological
considerations
applications.
Traditional
methodologies
require
estimation
probability
distribution
at
first.
show
step
introduce
bias
and,
consequently,
severely
impact
interpretations.
propose
parametric
method,
demonstrated
against
biases.
Second,
analysis,
there
variance
explain
proving
its
statistically
powerful
tracking
than
addition,
retains
characteristics
processing
when
complex
deep
networks.
NeuroImage,
Journal Year:
2023,
Volume and Issue:
272, P. 120040 - 120040
Published: March 17, 2023
During
listening,
brain
activity
tracks
the
rhythmic
structures
of
speech
signals.
Here,
we
directly
dissociated
contribution
neural
envelope
tracking
in
processing
acoustic
cues
from
that
related
to
linguistic
processing.
We
examined
changes
associated
with
comprehension
Noise-Vocoded
(NV)
using
magnetoencephalography
(MEG).
Participants
listened
NV
sentences
a
3-phase
training
paradigm:
(1)
pre-training,
where
stimuli
were
barely
comprehended,
(2)
exposure
original
clear
version
stimulus,
and
(3)
post-training,
same
gained
intelligibility
phase.
Using
this
paradigm,
tested
if
responses
signal
was
modulated
by
its
without
any
change
structure.
To
test
influence
spectral
degradation
on
independently
training,
participants
two
types
(4-band
2-band
speech),
but
only
trained
understand
4-band
speech.
Significant
observed
delta
range
relation
However,
failed
find
direct
effect
both
theta
ranges,
auditory
regions-of-interest
whole-brain
sensor-space
analyses.
This
suggests
acoustics
greatly
response
envelope,
caution
needs
be
taken
when
choosing
control
signals
for
speech-brain
analyses,
considering
slight
parameters
can
have
strong
effects
response.
PLoS Biology,
Journal Year:
2023,
Volume and Issue:
21(7), P. e3002178 - e3002178
Published: July 21, 2023
Speech
production
and
perception
are
fundamental
processes
of
human
cognition
that
both
rely
on
intricate
processing
mechanisms
still
poorly
understood.
Here,
we
study
these
by
using
magnetoencephalography
(MEG)
to
comprehensively
map
connectivity
regional
brain
activity
within
the
speech
envelope
during
continuous
speaking
listening.
Our
results
reveal
not
only
a
partly
shared
neural
substrate
for
but
also
dissociation
in
space,
delay,
frequency.
Neural
motor
frontal
areas
is
coupled
succeeding
delta
band
(1
3
Hz),
whereas
coupling
theta
range
follows
temporal
speaking.
showed
separation
bottom-up
top-down
signalling
distinct
frequency
bands
show
frequency-specific
channels
support
These
findings
further
shed
light
complex
interplay
between
different
regions
involved
perception.
Cerebral Cortex,
Journal Year:
2023,
Volume and Issue:
33(10), P. 6273 - 6281
Published: Jan. 10, 2023
Abstract
When
we
attentively
listen
to
an
individual’s
speech,
our
brain
activity
dynamically
aligns
the
incoming
acoustic
input
at
multiple
timescales.
Although
this
systematic
alignment
between
ongoing
and
speech
in
auditory
areas
is
well
established,
events
that
drive
phase-locking
are
not
fully
understood.
Here,
use
magnetoencephalographic
recordings
of
24
human
participants
(12
females)
while
they
were
listening
a
1
h
story.
We
show
whereas
speech–brain
coupling
associated
with
sustained
fluctuations
envelope
theta-frequency
range
(4–7
Hz),
tracking
low-frequency
delta
(below
Hz)
was
strongest
around
onsets
like
beginning
sentence.
Crucially,
bilateral
after
onsets,
proposing
during
continuous
perception
driven
by
onsets.
conclude
both
components
contribute
differentially
delta-
bands,
orchestrating
sampling
speech.
Thus,
results
suggest
temporal
dissociation
acoustically
oscillatory
tracking,
providing
valuable
implications
for
orchestration
time
scales.
Journal of Neural Engineering,
Journal Year:
2023,
Volume and Issue:
20(2), P. 026007 - 026007
Published: Feb. 22, 2023
Objective.The
human
brain
tracks
the
temporal
envelope
of
speech,
which
contains
essential
cues
for
speech
understanding.
Linear
models
are
most
common
tool
to
study
neural
tracking.
However,
information
on
how
is
processed
can
be
lost
since
nonlinear
relations
precluded.
Analysis
based
mutual
(MI),
other
hand,
detect
both
linear
and
gradually
becoming
more
popular
in
field
Yet,
several
different
approaches
calculating
MI
applied
with
no
consensus
approach
use.
Furthermore,
added
value
techniques
remains
a
subject
debate
field.
The
present
paper
aims
resolve
these
open
questions.Approach.We
analyzed
electroencephalography
(EEG)
data
participants
listening
continuous
analyses
models.Main
results.Comparing
approaches,
we
conclude
that
results
reliable
robust
using
Gaussian
copula
approach,
first
transforms
standard
Gaussians.
With
this
analysis
valid
technique
studying
Like
models,
it
allows
spatial
interpretations
processing,
peak
latency
analyses,
applications
multiple
EEG
channels
combined.
In
final
analysis,
tested
whether
components
were
response
by
removing
all
data.
We
robustly
detected
single-subject
level
analysis.Significance.We
demonstrate
processes
way.
Unlike
detects
such
relations,
proving
its
addition,
retains
characteristics
an
advantage
when
complex
(nonlinear)
deep
networks.
To
what
extent
does
speech
and
music
processing
rely
on
domain-specific
domain-general
neural
networks?
Using
whole-brain
intracranial
EEG
recordings
in
18
epilepsy
patients
listening
to
natural,
continuous
or
music,
we
investigated
the
presence
of
frequency-specific
network-level
brain
activity.
We
combined
it
with
a
statistical
approach
which
clear
operational
distinction
is
made
between
shared,
preferred,
domain-selective
responses.
show
that
majority
focal
activity
shared
processing.
Our
data
also
reveal
an
absence
anatomical
regional
selectivity.
Instead,
responses
are
restricted
distributed
coherent
oscillations,
typical
spectral
fingerprints.
work
highlights
importance
considering
natural
stimuli
dynamics
their
full
complexity
map
cognitive
functions.
To
what
extent
does
speech
and
music
processing
rely
on
domain-specific
domain-general
neural
networks?
Using
whole-brain
intracranial
EEG
recordings
in
18
epilepsy
patients
listening
to
natural,
continuous
or
music,
we
investigated
the
presence
of
frequency-specific
network-level
brain
activity.
We
combined
it
with
a
statistical
approach
which
clear
operational
distinction
is
made
between
shared
,
preferred,
domain-
selective
responses.
show
that
majority
focal
activity
processing.
Our
data
also
reveal
an
absence
anatomical
regional
selectivity.
Instead,
domain-selective
responses
are
restricted
distributed
coherent
oscillations,
typical
spectral
fingerprints.
work
highlights
importance
considering
natural
stimuli
dynamics
their
full
complexity
map
cognitive
functions.
eNeuro,
Journal Year:
2025,
Volume and Issue:
unknown, P. ENEURO.0368 - 24.2024
Published: Jan. 16, 2025
Observing
lip
movements
of
a
speaker
facilitates
speech
understanding,
especially
in
challenging
listening
situations.
Converging
evidence
from
neuroscientific
studies
shows
stronger
neural
responses
to
audiovisual
stimuli
compared
audio-only
stimuli.
However,
the
interindividual
variability
this
contribution
movement
information
and
its
consequences
on
behavior
are
unknown.
We
analyzed
source-localized
magnetoencephalographic
(MEG)
29
normal-hearing
participants
(12
female)
speech,
both
with
without
wearing
surgical
face
mask,
presence
or
absence
distractor
speaker.
Using
temporal
response
functions
(TRFs)
quantify
tracking,
we
show
that
are,
general,
enhanced
when
is
challenging.
After
controlling
for
acoustics,
contribute
particularly
present.
extent
visual
tracking
varied
greatly
among
participants.
Probing
behavioral
relevance,
demonstrate
individuals
who
higher
terms
drop
comprehension
an
increase
perceived
difficulty
mouth
occluded
by
mask.
By
contrast,
no
effect
was
found
not
occluded.
provide
novel
insights
how
varies
revealing
negative
absent.
Our
results
also
offer
potential
implications
objective
assessments
perception.
Significance
Statement
In
complex
auditory
environments,
simultaneous
conversations
pose
challenge
comprehension.
investigated
level,
aid
such
situations
what
observing
enhances
rely
more
deterioration
wears
Remarkably,
case
mask
worn
findings
reveal
differences
applications
Current Biology,
Journal Year:
2024,
Volume and Issue:
34(15), P. 3537 - 3549.e5
Published: July 23, 2024
Decoding
human
speech
requires
the
brain
to
segment
incoming
acoustic
signal
into
meaningful
linguistic
units,
ranging
from
syllables
and
words
phrases.
Integrating
these
constituents
a
coherent
percept
sets
root
of
compositional
meaning
hence
understanding.
One
important
cue
for
segmentation
in
natural
is
prosodic
cues,
such
as
pauses,
but
their
interplay
with
higher-level
processing
still
unknown.
Here,
we
dissociate
neural
tracking
pauses
multi-word
chunks
using
magnetoencephalography
(MEG).
We
find
that
manipulating
regularity
disrupts
slow
speech-brain
bilaterally
auditory
areas
(below
2
Hz)
turn
increases
left-lateralized
coherence
higher-frequency
activity
at
onsets
(around
25-45
Hz).
Critically,
also
chunks-defined
short,
bundles
inter-word
dependencies-are
processed
through
rhythmic
fluctuations
low-frequency
independently
cues.
Importantly,
alignment
chunk
accuracy
an
encoding
model
bilateral
frontal
while
controlling
effect
acoustics.
Our
findings
provide
novel
insights
basis
perception,
demonstrating
both
features
(prosodic
cues)
abstract
timescale
are
underpinned
by
electrophysiological
delta
frequency
range.
NeuroImage,
Journal Year:
2023,
Volume and Issue:
284, P. 120424 - 120424
Published: Oct. 30, 2023
Magnetoencephalography
and
electroencephalography
(M/EEG)
seed-based
connectivity
analysis
requires
the
extraction
of
measures
from
regions
interest
(ROI).
M/EEG
ROI-derived
source
activity
can
be
treated
in
different
ways.
It
is
possible,
for
instance,
to
average
each
ROI's
time
series
prior
calculating
measures.
Alternatively,
one
compute
maps
element
ROI
dimensionality
reduction
obtain
a
single
map.
The
impact
these
strategies
on
results
still
unclear.
Here,
we
address
this
question
within
large
MEG
resting
state
cohort
(N=113)
simulated
data.
We
consider
68
ROIs
(Desikan-Kiliany
atlas),
two
(phase
locking
value-PLV,
its
imaginary
counterpart-
ciPLV),
three
frequency
bands
(theta
4-8
Hz,
alpha
9-12
beta
15-30
Hz).
compare
four
methods:
(i)
mean,
or
(ii)
PCA
seed
before
computing
connectivity,
map
(iii)
average,
(iv)
maximum
after
seed.
Hierarchical
clustering
then
applied
outputs
across
multiple
strategies,
followed
by
direct
contrasts
methods.
Finally,
are
validated
using
set
realistic
simulations.
show
that
ROI-based
vary
remarkably
terms
magnitude
spatial
distribution.
Dimensionality
procedures
conducted
more
similar
each-other,
while
approach
most
dissimilar
other
approaches.
Although
differences
methods
consistent
bands,
they
influenced
metric
size.
Greater
were
observed
ciPLV
than
PLV,
larger
ROIs.
Realistic
simulations
confirmed
aggregation
generally
accurate
but
have
lower
specificity
(higher
rate
false
positive
connections).
Though
computationally
demanding,
should
preferred
when
higher
sensitivity
desired.
Given
remarkable
procedures,
caution
warranted
comparing
studies
applying
bioRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2023,
Volume and Issue:
unknown
Published: April 18, 2023
Abstract
Observing
lip
movements
of
a
speaker
is
known
to
facilitate
speech
understanding,
especially
in
challenging
listening
situations.
Converging
evidence
from
neuroscientific
studies
shows
enhanced
processing
audiovisual
stimuli.
However,
the
interindividual
variability
this
visual
benefit
and
its
consequences
on
behavior
are
unknown.
Here,
we
analyzed
source-localized
magnetoencephalographic
(MEG)
responses
normal-hearing
participants
with
or
without
an
additional
distractor
speaker.
Using
temporal
response
functions
(TRFs),
show
that
neural
are,
general,
when
challenging.
After
conducting
crucial
control
for
acoustics,
effectively
contribute
higher
tracking,
particularly
present.
extent
varied
greatly
among
participants.
Probing
behavioral
relevance,
individuals
who
more
movement
information
terms
stronger
drop
performance
increase
perceived
difficulty
mouth
occluded
by
surgical
face
mask.
By
contrast,
no
effect
was
found
not
occluded.
We
provide
novel
insights
how
tracking
varies
individuals.
Furthermore,
reveal
relevance
demonstrating
negative
absent.
Our
results
also
offer
potential
implications
future
objective
assessments
perception.