Our
movements
result
in
predictable
sensory
feedback
that
is
often
multimodal.
Based
on
deviations
between
predictions
and
actual
input,
primary
areas
of
cortex
have
been
shown
to
compute
sensorimotor
prediction
errors.
How
errors
one
modality
influence
the
computation
another
still
unclear.
To
investigate
multimodal
mouse
auditory
(ACx),
we
used
a
virtual
environment
experimentally
couple
running
both
self-generated
visual
feedback.
Using
two-photon
microscopy,
first
characterized
responses
layer
2/3
(L2/3)
neurons
sounds,
stimuli,
onsets
found
all
three
stimuli.
Probing
evoked
by
audiomotor
mismatches,
they
closely
resemble
visuomotor
mismatch
(V1).
Finally,
testing
for
cross
modal
coupling
sound
amplitude
flow
speed
running,
were
amplified
when
paired
with
concurrent
mismatches.
results
demonstrate
non-hierarchical
interactions
shape
error
cortical
L2/3.
***
Dear
reader,
please
note
this
manuscript
formatted
standard
submission
format,
statistical
information
provided
Table
S1.
Cell Reports,
Journal Year:
2025,
Volume and Issue:
44(4), P. 115538 - 115538
Published: April 1, 2025
Under
the
predictive
coding
framework,
brain
generates
a
model
of
environment
based
on
previous
experiences.
Incoming
sensory
information
is
compared
to
this
model,
such
that
if
predictions
do
not
match
inputs,
prediction
error
generated.
Predictions
are
passed
top-down,
and
errors
emerge
when
bottom-up
does
predictions.
Prediction
occur
sequentially
in
primary
auditory
cortex
(A1)
then
medial
prefrontal
(mPFC).
Here,
we
test
hypothesis
mPFC
sends
contribute
generation
errors.
We
used
optogenetics
block
top-down
signals
from
while
recording
neuronal
A1
under
classical
"oddball"
paradigm.
Blocking
reduces
response
rare
sounds,
it
affect
responses
predictable
or
random
sounds.
Our
results
provide
empirical
evidence
for
enhance
unpredicted
stimuli.
The
brain
predicts
regularities
in
sensory
inputs
at
multiple
complexity
levels,
with
neuronal
mechanisms
that
remain
elusive.
Here,
we
monitored
auditory
cortex
activity
during
the
local-global
paradigm,
a
protocol
nesting
different
regularity
levels
sound
sequences.
We
observed
mice
encode
local
predictions
based
on
stimulus
occurrence
and
transition
probabilities,
because
responses
are
boosted
upon
prediction
violation.
This
boosting
was
due
to
both
short-term
adaptation
an
adaptation-independent
surprise
mechanism
resisting
anesthesia.
In
parallel,
only
wakefulness,
VIP
interneurons
responded
omission
of
locally
expected
repeat
sequence
ending,
thus
providing
chunking
signal
potentially
useful
for
establishing
global
structure.
When
this
structure
violated,
by
either
shortening
or
ending
it
but
globally
unexpected
transition,
slightly
increased
PV
neurons,
respectively.
Hence,
distinct
cellular
predict
bioRxiv (Cold Spring Harbor Laboratory),
Journal Year:
2023,
Volume and Issue:
unknown
Published: Jan. 7, 2023
Comparing
expectation
with
experience
is
an
important
neural
computation
performed
throughout
the
brain
and
a
hallmark
of
predictive
processing.
Experiments
that
alter
sensory
outcome
animal's
behavior
reveal
enhanced
responses
to
unexpected
self-generated
stimuli,
indicating
populations
neurons
in
cortex
may
reflect
prediction
errors
-
mismatches
between
experience.
However,
stimuli
could
also
arise
through
non-predictive
mechanisms,
such
as
movement-based
facilitation
neuron's
inherent
sound
responses.
If
error
exist
cortex,
it
unknown
whether
they
manifest
general
responses,
or
respond
specificity
distinct
stimulus
dimensions.
To
answer
these
questions,
we
trained
mice
expect
simple
sound-generating
recorded
auditory
activity
heard
either
expected
sounds
deviated
from
one
multiple
Our
data
learns
suppress
along
acoustic
dimensions
simultaneously.
We
identify
population
are
not
responsive
passive
but
explicitly
encode
errors.
These
abundant
only
animals
learned
motor-sensory
expectation,
two
specific
violations
rather
than
generic
signal.
Our
movements
result
in
predictable
sensory
feedback
that
is
often
multimodal.
Based
on
deviations
between
predictions
and
actual
input,
primary
areas
of
cortex
have
been
shown
to
compute
sensorimotor
prediction
errors.
How
errors
one
modality
influence
the
computation
another
still
unclear.
To
investigate
multimodal
mouse
auditory
(ACx),
we
used
a
virtual
environment
experimentally
couple
running
both
self-generated
visual
feedback.
Using
two-photon
microscopy,
first
characterized
responses
layer
2/3
(L2/3)
neurons
sounds,
stimuli,
onsets
found
all
three
stimuli.
Probing
evoked
by
audiomotor
mismatches,
they
closely
resemble
visuomotor
mismatch
(V1).
Finally,
testing
for
cross
modal
coupling
sound
amplitude
flow
speed
running,
were
amplified
when
paired
with
concurrent
mismatches.
results
demonstrate
non-hierarchical
interactions
shape
error
cortical
L2/3.
Our
movements
result
in
predictable
sensory
feedback
that
is
often
multimodal.
Based
on
deviations
between
predictions
and
actual
input,
primary
areas
of
cortex
have
been
shown
to
compute
sensorimotor
prediction
errors.
How
errors
one
modality
influence
the
computation
another
still
unclear.
To
investigate
multimodal
mouse
auditory
(ACx),
we
used
a
virtual
environment
experimentally
couple
running
both
self-generated
visual
feedback.
Using
two-photon
microscopy,
first
characterized
responses
layer
2/3
(L2/3)
neurons
sounds,
stimuli,
onsets
found
all
three
stimuli.
Probing
evoked
by
audiomotor
mismatches,
they
closely
resemble
visuomotor
mismatch
(V1).
Finally,
testing
for
cross
modal
coupling
sound
amplitude
flow
speed
running,
were
amplified
when
paired
with
concurrent
mismatches.
results
demonstrate
non-hierarchical
interactions
shape
error
cortical
L2/3.
***
Dear
reader,
please
note
this
manuscript
formatted
standard
submission
format,
statistical
information
provided
Table
S1.