bioRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2022,
Volume and Issue:
unknown
Published: April 8, 2022
ABSTRACT
Spoken
language
comprehension
requires
rapid
and
continuous
integration
of
information,
from
lower-level
acoustic
to
higher-level
linguistic
features.
Much
this
processing
occurs
in
the
cerebral
cortex.
Its
neural
activity
exhibits,
for
instance,
correlates
predictive
processing,
emerging
at
delays
a
few
hundred
milliseconds.
However,
auditory
pathways
are
also
characterized
by
extensive
feedback
loops
cortical
areas
ones
as
well
subcortical
structures.
Early
can
therefore
be
influenced
cognitive
processes,
but
it
remains
unclear
whether
such
contributes
processing.
Here,
we
investigated
early
speech-evoked
that
emerges
fundamental
frequency.
We
analyzed
EEG
recordings
obtained
when
subjects
listened
story
read
single
speaker.
identified
response
tracking
speaker’s
frequency
occurred
delay
11
ms,
while
another
elicited
high-frequency
modulation
envelope
higher
harmonics
exhibited
larger
magnitude
longer
latency
about
18
ms.
Subsequently,
determined
these
responses
each
individual
word
story.
then
quantified
context-independent
used
model
compute
context-dependent
surprisal
precision.
The
represented
how
predictable
is,
given
previous
context,
precision
reflected
confidence
predicting
next
past
context.
found
word-level
were
predominantly
features:
average
its
variability.
Amongst
features,
only
showed
weak
significant
modulation.
Our
results
show
is
already
suggesting
top-down
response.
PLoS ONE,
Journal Year:
2024,
Volume and Issue:
19(2), P. e0297826 - e0297826
Published: Feb. 8, 2024
Perception
of
sounds
and
speech
involves
structures
in
the
auditory
brainstem
that
rapidly
process
ongoing
stimuli.
The
role
these
processing
can
be
investigated
by
measuring
their
electrical
activity
using
scalp-mounted
electrodes.
However,
typical
analysis
methods
involve
averaging
neural
responses
to
many
short
repetitive
stimuli
bear
little
relevance
daily
listening
environments.
Recently,
subcortical
more
ecologically
relevant
continuous
were
detected
linear
encoding
models.
These
estimate
temporal
response
function
(TRF),
which
is
a
regression
model
minimises
error
between
measured
signal
predictor
derived
from
stimulus.
Using
predictors
highly
non-linear
peripheral
system
may
improve
TRF
estimation
accuracy
peak
detection.
Here,
we
compare
both
simple
complex
models
for
estimating
TRFs
on
electroencephalography
(EEG)
data
24
participants
speech.
We
also
investigate
length
required
TRFs,
find
around
12
minutes
sufficient
clear
wave
V
peaks
(>3
dB
SNR)
seen
nearly
all
participants.
Interestingly,
filterbank-based
yield
SNRs
are
not
significantly
different
those
estimated
nerve,
provided
nonlinear
effects
adaptation
appropriately
modelled.
Crucially,
computing
simpler
than
50
times
faster
compared
model.
This
work
paves
way
efficient
modelling
detection
speech,
lead
improved
diagnosis
metrics
hearing
impairment
assistive
technology.
Scientific Reports,
Journal Year:
2024,
Volume and Issue:
14(1)
Published: Jan. 8, 2024
Music
and
speech
are
encountered
daily
unique
to
human
beings.
Both
transformed
by
the
auditory
pathway
from
an
initial
acoustical
encoding
higher
level
cognition.
Studies
of
cortex
have
revealed
distinct
brain
responses
music
speech,
but
differences
may
emerge
in
or
be
inherited
different
subcortical
encoding.
In
first
part
this
study,
we
derived
brainstem
response
(ABR),
a
measure
encoding,
recorded
using
two
analysis
methods.
The
method,
described
previously
acoustically
based,
yielded
very
ABRs
between
sound
classes.
second
however,
developed
here
based
on
physiological
model
periphery,
gave
highly
correlated
speech.
We
determined
superiority
method
through
several
metrics,
suggesting
there
is
no
appreciable
impact
stimulus
class
(i.e.,
vs
speech)
way
acoustics
encoded
subcortically.
study's
part,
considered
cortex.
Our
new
resulted
cortical
becoming
more
similar
with
remaining
differences.
results
taken
together
suggest
that
evidence
for
stimulus-class
dependent
processing
at
not
level.
Trends in Hearing,
Journal Year:
2024,
Volume and Issue:
28
Published: Jan. 1, 2024
The
auditory
brainstem
response
(ABR)
is
a
valuable
clinical
tool
for
objective
hearing
assessment,
which
conventionally
detected
by
averaging
neural
responses
to
thousands
of
short
stimuli.
Progressing
beyond
these
unnatural
stimuli,
continuous
speech
presented
via
earphones
have
been
recently
using
linear
temporal
functions
(TRFs).
Here,
we
extend
earlier
studies
measuring
subcortical
in
the
sound-field,
and
assess
amount
data
needed
estimate
TRFs.
Electroencephalography
(EEG)
was
recorded
from
24
normal
participants
while
they
listened
clicks
stories
loudspeakers.
Subcortical
TRFs
were
computed
after
accounting
non-linear
processing
periphery
either
stimulus
rectification
or
an
nerve
model.
Our
results
demonstrated
that
could
be
reliably
measured
sound-field.
estimated
models
outperformed
simple
rectification,
16
minutes
sufficient
all
show
clear
wave
V
peaks
both
sound-field
highly
consistent
earphone
conditions,
with
click
ABRs.
However,
required
slightly
more
(16
minutes)
achieve
compared
(12
minutes),
possibly
due
effects
room
acoustics.
By
investigating
this
study
lays
groundwork
bringing
assessment
closer
real-life
may
lead
improved
evaluations
smart
technologies.
Frontiers in Human Neuroscience,
Journal Year:
2025,
Volume and Issue:
19
Published: April 9, 2025
The
audio-visual
benefit
in
speech
perception-where
congruent
visual
input
enhances
auditory
processing-is
well-documented
across
age
groups,
particularly
challenging
listening
conditions
and
among
individuals
with
varying
hearing
abilities.
However,
most
studies
rely
on
highly
controlled
laboratory
environments
scripted
stimuli.
Here,
we
examine
the
using
unscripted,
natural
from
untrained
speakers
within
a
virtual
acoustic
environment.
Using
electroencephalography
(EEG)
cortical
tracking,
assessed
neural
responses
audio-visual,
audio-only,
visual-only,
masked-lip
to
isolate
role
of
lip
movements.
Additionally,
analysed
individual
differences
features
speakers,
including
pitch,
jitter,
lip-openness,
explore
their
influence
tracking
benefit.
Results
showed
significant
enhancement
background
noise,
condition
performing
similarly
audio-only
condition,
emphasizing
importance
movements
adverse
situations.
Our
findings
reveal
feasibility
naturalistic
stimuli
underscore
impact
speaker
characteristics
integration
real-world
contexts.
Frontiers in Neuroscience,
Journal Year:
2022,
Volume and Issue:
16
Published: July 22, 2022
Spoken
language
comprehension
requires
rapid
and
continuous
integration
of
information,
from
lower-level
acoustic
to
higher-level
linguistic
features.
Much
this
processing
occurs
in
the
cerebral
cortex.
Its
neural
activity
exhibits,
for
instance,
correlates
predictive
processing,
emerging
at
delays
a
few
100
ms.
However,
auditory
pathways
are
also
characterized
by
extensive
feedback
loops
cortical
areas
ones
as
well
subcortical
structures.
Early
can
therefore
be
influenced
cognitive
processes,
but
it
remains
unclear
whether
such
contributes
processing.
Here,
we
investigated
early
speech-evoked
that
emerges
fundamental
frequency.
We
analyzed
EEG
recordings
obtained
when
subjects
listened
story
read
single
speaker.
identified
response
tracking
speaker's
frequency
occurred
delay
11
ms,
while
another
elicited
high-frequency
modulation
envelope
higher
harmonics
exhibited
larger
magnitude
longer
latency
about
18
ms
with
an
additional
significant
component
around
40
Notably,
earlier
components
likely
originate
structures,
latter
presumably
involves
contributions
regions.
Subsequently,
determined
these
responses
each
individual
word
story.
then
quantified
context-independent
used
model
compute
context-dependent
surprisal
precision.
The
represented
how
predictable
is,
given
previous
context,
precision
reflected
confidence
predicting
next
past
context.
found
word-level
were
predominantly
features:
average
its
variability.
Amongst
features,
only
showed
weak
modulation.
Our
results
show
is
already
suggesting
top-down
response.
eNeuro,
Journal Year:
2024,
Volume and Issue:
unknown, P. ENEURO.0545 - 23.2024
Published: April 24, 2024
More
and
more
patients
worldwide
are
diagnosed
with
dementia,
which
emphasizes
the
urgent
need
for
early
detection
markers.
In
this
study,
we
built
on
auditory
hypersensitivity
theory
of
a
previous
study—which
postulated
that
responses
to
input
in
subcortex
as
well
cortex
enhanced
cognitive
decline—,
examined
encoding
natural
continuous
speech
at
both
neural
levels
its
indicative
potential
decline.
We
recruited
study
participants
aged
60
years
older,
who
were
divided
into
two
groups
based
Montreal
Cognitive
Assessment,
one
group
low
scores
(n
=
19,
signs
decline),
control
25).
Participants
completed
an
audiometric
assessment
then
recorded
their
electroencephalography
while
they
listened
audiobook
click
sounds.
derived
Temporal
Response
Functions
evoked
potentials
from
data
response
amplitudes
predict
decline,
controlling
age
hearing
loss.
Contrary
our
expectations,
no
evidence
was
observed
decline;
comparable
groups.
Moreover,
combination
showed
predictive
value
These
results
challenge
proposed
hypothesis
emphasize
further
research
identify
reliable
markers
Significance
statement
Research
decline
needs
studies
uncovering
subtle
changes
pre-dementia
stage.
Neural
processing
have
has
not
been
widely
explored
studies.
Here,
first
time,
used
natural,
continuously
spoken
language
examine
older
adults
without
quantified
how
brain
tracks
only
cortical
but
also
subcortical
level.
contrast
suggesting
found
differences.
believe
represents
important
contribution
search
health
old
age.
Scientific Reports,
Journal Year:
2024,
Volume and Issue:
14(1)
Published: Aug. 17, 2024
The
multivariate
temporal
response
function
(mTRF)
is
an
effective
tool
for
investigating
the
neural
encoding
of
acoustic
and
complex
linguistic
features
in
natural
continuous
speech.
In
this
study,
we
investigated
how
representations
speech
derived
from
stimuli
are
related
to
early
signs
cognitive
decline
older
adults,
taking
into
account
effects
hearing.
Participants
without
(
bioRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2024,
Volume and Issue:
unknown
Published: May 30, 2024
Abstract
Several
tools
have
recently
been
developed
to
derive
the
Auditory
Brainstem
Response
(ABR)
from
continuous
natural
speech,
facilitating
investigation
into
subcortical
encoding
of
speech.
These
rely
on
deconvolution,
which
models
auditory
pathway
as
a
linear
system,
where
nonlinearly
processed
stimulus
is
taken
input
(i.e.,
regressor),
electroencephalogram
(EEG)
data
output,
and
ABR
impulse
response
deconvolved
recorded
EEG
regressor.
In
this
study,
we
analyzed
recordings
subjects
listening
both
unaltered
speech
synthesized
“peaky
speech.”
We
compared
derived
ABRs
using
three
regressors:
half-wave
rectified
(HWR)
Maddox
Lee
(2018),
glottal
pulse
train
(GP)
Polonenko
(2021),
nerve
modeled
(ANM)
Shan
et
al.
(2024).
Our
evaluation
focused
fidelity,
efficiency,
practicality
each
method
in
different
scenarios.
The
results
indicate
that
ANM
regressor
for
peaky
GP
provided
best
performance,
whereas
HWR
demonstrated
relatively
poorer
performance.
findings
study
will
guide
future
research
selecting
most
appropriate
paradigm
derivation
continuous,
naturalistic
eNeuro,
Journal Year:
2024,
Volume and Issue:
11(8), P. ENEURO.0135 - 24.2024
Published: Aug. 1, 2024
The
auditory
brainstem
response
(ABR)
is
a
measure
of
subcortical
activity
in
to
stimuli.
wave
V
peak
the
ABR
depends
on
stimulus
intensity
level,
and
has
been
widely
used
for
clinical
hearing
assessment.
Conventional
methods
estimate
average
electroencephalography
(EEG)
responses
short
unnatural
stimuli
such
as
clicks.
Recent
work
moved
toward
more
ecologically
relevant
continuous
speech
using
linear
deconvolution
models
called
temporal
functions
(TRFs).
Investigating
whether
TRF
waveform
changes
with
crucial
step
use
natural
assessments
involving
responses.
Here,
we
develop
level-dependent
TRFs
EEG
data
collected
from
21
participants
listening
presented
at
4
different
levels.
We
find
that
can
be
detected
almost
all
participants,
are
consistent
click-ABR
V.
also
investigate
most
suitable
peripheral
model
generate
predictors
simple
gammatone
filterbanks
perform
best.
Additionally,
around
6
min
may
sufficient
detecting
effects
peaks
above
noise
floor
segments
higher
intensity.
Finally,
show
proof-of-concept
even
inherent
fluctuations
speech.