Dataset size considerations for robust acoustic and phonetic speech encoding models in EEG
Frontiers in Human Neuroscience,
Journal Year:
2023,
Volume and Issue:
16
Published: Jan. 20, 2023
In
many
experiments
that
investigate
auditory
and
speech
processing
in
the
brain
using
electroencephalography
(EEG),
experimental
paradigm
is
often
lengthy
tedious.
Typically,
experimenter
errs
on
side
of
including
more
data,
trials,
therefore
conducting
a
longer
task
to
ensure
data
are
robust
effects
measurable.
Recent
studies
used
naturalistic
stimuli
brain's
response
individual
or
combination
multiple
features
system
identification
techniques,
such
as
multivariate
temporal
receptive
field
(mTRF)
analyses.
The
neural
collected
from
must
be
divided
into
training
set
test
fit
validate
mTRF
weights.
While
good
strategy
clearly
collect
much
feasible,
it
unclear
how
needed
achieve
stable
results.
Furthermore,
whether
specific
stimulus
for
fitting
choice
feature
representation
affects
would
required
generalizable
Here,
we
previously
EEG
our
lab
sentence
movie
well
an
open-source
dataset
audiobook
better
understand
needs
measuring
acoustic
phonetic
tuning.
We
found
structure
tested
here
stabilizes
after
collecting
approximately
200
s
TIMIT
sentences,
around
600
trailers
460
data.
Thus,
provide
suggestions
minimum
amount
necessary
mTRFs
listening
Our
findings
motivated
by
highly
practical
concerns
when
working
with
children,
patient
populations,
others
who
may
not
tolerate
long
study
sessions.
These
will
aid
future
researchers
wish
healthy
clinical
populations
while
minimizing
participant
fatigue
retaining
signal
quality.
Language: Английский
Cortical Responses Time-Locked to Continuous Speech in the High-Gamma Band Depend on Selective Attention
bioRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2023,
Volume and Issue:
unknown
Published: July 24, 2023
Auditory
cortical
responses
to
speech
obtained
by
magnetoencephalography
(MEG)
show
robust
tracking
the
speaker's
fundamental
frequency
in
high-gamma
band
(70-200
Hz),
but
little
is
currently
known
about
whether
such
depend
on
focus
of
selective
attention.
In
this
study
22
human
subjects
listened
concurrent,
fixed-rate,
from
male
and
female
speakers,
were
asked
selectively
attend
one
speaker
at
a
time,
while
their
neural
recorded
with
MEG.
The
pitch
range
coincided
lower
band,
whereas
higher
had
much
less
overlap,
only
upper
end
band.
Neural
analyzed
using
temporal
response
function
(TRF)
framework.
As
expected,
demonstrate
male's
speech,
peak
latency
approximately
40
ms.
Critically,
magnitude
depends
attention:
significantly
greater
when
attended
than
it
not
attended,
under
acoustically
identical
conditions.
This
clear
demonstration
that
even
very
early
auditory
are
influenced
top-down,
cognitive,
processing
mechanisms.
Language: Английский