bioRxiv (Cold Spring Harbor Laboratory),
Год журнала:
2023,
Номер
unknown
Опубликована: Ноя. 1, 2023
SUMMARY
Our
movements
result
in
predictable
sensory
feedback
that
is
often
multimodal.
Based
on
deviations
between
predictions
and
actual
input,
primary
areas
of
cortex
have
been
shown
to
compute
sensorimotor
prediction
errors.
How
errors
one
modality
influence
the
computation
another
still
unclear.
To
investigate
multimodal
mouse
auditory
(ACx),
we
used
a
virtual
environment
experimentally
couple
running
both
self-generated
visual
feedback.
Using
two-photon
microscopy,
first
characterized
responses
layer
2/3
(L2/3)
neurons
sounds,
stimuli,
onsets
found
all
three
stimuli.
Probing
evoked
by
audiomotor
mismatches,
they
closely
resemble
visuomotor
mismatch
(V1).
Finally,
testing
for
cross
modal
coupling
sound
amplitude
flow
speed
running,
were
amplified
when
paired
with
concurrent
mismatches.
results
demonstrate
non-hierarchical
interactions
shape
error
cortical
L2/3.
Acetylcholine
is
released
in
visual
cortex
by
axonal
projections
from
the
basal
forebrain.
The
signals
conveyed
these
and
their
computational
significance
are
still
unclear.
Using
two-photon
calcium
imaging
behaving
mice,
we
show
that
forebrain
cholinergic
axons
mouse
provide
a
binary
locomotion
state
signal.
In
axons,
found
no
evidence
of
responses
to
stimuli
or
visuomotor
prediction
errors.
While
optogenetic
activation
isolation
did
not
drive
local
neuronal
activity,
when
paired
with
stimuli,
it
resulted
layer-specific
increases
activity.
Responses
layer
5
neurons
both
top-down
bottom-up
inputs
were
increased
amplitude
decreased
latency,
whereas
those
2/3
remained
unchanged.
opto-
chemogenetic
manipulations
acetylcholine
underlie
locomotion-associated
decorrelation
activity
between
5.
Our
results
suggest
augments
responsiveness
outside
network,
possibly
enabling
faster
switching
internal
representations
during
locomotion.
Nature Communications,
Год журнала:
2024,
Номер
15(1)
Опубликована: Июль 30, 2024
Time
courses
of
neural
responses
underlie
real-time
sensory
processing
and
perception.
How
these
temporal
dynamics
change
may
be
fundamental
to
how
systems
adapt
different
perceptual
demands.
By
simultaneously
recording
from
hundreds
neurons
in
mouse
primary
visual
cortex,
we
examined
population
stimuli
at
sub-second
timescales,
during
behavioural
states.
We
discovered
that
active
states
characterised
by
locomotion,
single-neurons
shift
transient
sustained
response
modes,
facilitating
rapid
emergence
stimulus
tuning.
Differences
single-neuron
were
associated
with
changes
correlations,
including
faster
stabilisation
stimulus-evoked
the
structure
correlations
locomotion.
Using
Factor
Analysis,
latent
trajectories
activity
make
more
direct
transitions
between
baseline
stimulus-encoding
This
could
partly
explained
dampening
oscillatory
present
stationary
Functionally,
collectively
enabled
faster,
stable
efficient
encoding
new
information
These
findings
reveal
a
principle
demands,
where
flexible
govern
speed
stability
encoding.
The
last
few
years
have
seen
an
explosion
in
the
number
of
tools
available
to
measure
neuronal
activity
using
fluorescence
imaging
(Chen
et
al.,
2013;
Feng
2019;
Jing
Sun
2018;
Wan
2021).
When
performed
vivo,
these
measurements
are
invariably
contaminated
by
hemodynamic
occlusion
artifacts.
In
widefield
calcium
imaging,
this
problem
is
well
recognized.
For
two-photon
however,
effects
only
been
sparsely
characterized.
Here
we
perform
a
quantification
changes
observed
with
GFP
expression
both
and
imaging.
We
find
that
many
instances
magnitude
signal
attributable
comparable
sensors.
Moreover,
were
spatially
heterogeneous,
over
cortical
regions
across
depth,
exhibited
complex
relationship
behavior.
Thus,
important
caveat
consider
when
analyzing
interpreting
not
just
but
also
data.
The
last
few
years
have
seen
an
explosion
in
the
number
of
tools
available
to
measure
neuronal
activity
using
fluorescence
imaging
(Chen
et
al.,
2013;
Feng
2019;
Jing
Sun
2018;
Wan
2021).
When
performed
vivo,
these
measurements
are
invariably
contaminated
by
hemodynamic
occlusion
artifacts.
In
widefield
calcium
imaging,
this
problem
is
well
recognized.
For
two-photon
however,
effects
only
been
sparsely
characterized.
Here
we
perform
a
quantification
changes
observed
with
GFP
expression
both
and
imaging.
We
find
that
many
instances
magnitude
signal
attributable
comparable
sensors.
Moreover,
were
spatially
heterogeneous,
over
cortical
regions
across
depth,
exhibited
complex
relationship
behavior.
Thus,
important
caveat
consider
when
analyzing
interpreting
not
just
but
also
data.
Our
movements
result
in
predictable
sensory
feedback
that
is
often
multimodal.
Based
on
deviations
between
predictions
and
actual
input,
primary
areas
of
cortex
have
been
shown
to
compute
sensorimotor
prediction
errors.
How
errors
one
modality
influence
the
computation
another
still
unclear.
To
investigate
multimodal
mouse
auditory
(ACx),
we
used
a
virtual
environment
experimentally
couple
running
both
self-generated
visual
feedback.
Using
two-photon
microscopy,
first
characterized
responses
layer
2/3
(L2/3)
neurons
sounds,
stimuli,
onsets
found
all
three
stimuli.
Probing
evoked
by
audiomotor
mismatches,
they
closely
resemble
visuomotor
mismatch
(V1).
Finally,
testing
for
cross
modal
coupling
sound
amplitude
flow
speed
running,
were
amplified
when
paired
with
concurrent
mismatches.
results
demonstrate
non-hierarchical
interactions
shape
error
cortical
L2/3.
Our
movements
result
in
predictable
sensory
feedback
that
is
often
multimodal.
Based
on
deviations
between
predictions
and
actual
input,
primary
areas
of
cortex
have
been
shown
to
compute
sensorimotor
prediction
errors.
How
errors
one
modality
influence
the
computation
another
still
unclear.
To
investigate
multimodal
mouse
auditory
cortex,
we
used
a
virtual
environment
experimentally
couple
running
both
self-generated
visual
feedback.
Using
two-photon
microscopy,
first
characterized
responses
layer
2/3
(L2/3)
neurons
sounds,
stimuli,
onsets
found
all
three
stimuli.
Probing
evoked
by
audiomotor
(AM)
mismatches,
they
closely
resemble
visuomotor
(VM)
mismatch
(V1).
Finally,
testing
for
cross
modal
AM
coupling
sound
amplitude
flow
speed
running,
were
amplified
when
paired
with
concurrent
VM
mismatches.
results
demonstrate
non-hierarchical
interactions
shape
error
cortical
L2/3.
The
last
few
years
have
seen
an
explosion
in
the
number
of
tools
available
to
measure
neuronal
activity
using
fluorescence
imaging
(Chen
et
al.,
2013;
Feng
2019;
Jing
Sun
2018;
Wan
2021).
When
performed
vivo,
these
measurements
are
invariably
contaminated
by
hemodynamic
occlusion
artifacts.
In
widefield
calcium
imaging,
this
problem
is
well
recognized.
For
two-photon
however,
effects
only
been
sparsely
characterized.
Here
we
perform
a
quantification
changes
observed
with
GFP
expression
both
and
imaging.
We
find
that
many
instances
magnitude
signal
attributable
comparable
sensors.
Moreover,
were
spatially
heterogeneous,
over
cortical
regions
across
depth,
exhibited
complex
relationship
behavior.
Thus,
important
caveat
consider
when
analyzing
interpreting
not
just
but
also
data.
Our
movements
result
in
predictable
sensory
feedback
that
is
often
multimodal.
Based
on
deviations
between
predictions
and
actual
input,
primary
areas
of
cortex
have
been
shown
to
compute
sensorimotor
prediction
errors.
How
errors
one
modality
influence
the
computation
another
still
unclear.
To
investigate
multimodal
mouse
auditory
(ACx),
we
used
a
virtual
environment
experimentally
couple
running
both
self-generated
visual
feedback.
Using
two-photon
microscopy,
first
characterized
responses
layer
2/3
(L2/3)
neurons
sounds,
stimuli,
onsets
found
all
three
stimuli.
Probing
evoked
by
audiomotor
mismatches,
they
closely
resemble
visuomotor
mismatch
(V1).
Finally,
testing
for
cross
modal
coupling
sound
amplitude
flow
speed
running,
were
amplified
when
paired
with
concurrent
mismatches.
results
demonstrate
non-hierarchical
interactions
shape
error
cortical
L2/3.
Our
movements
result
in
predictable
sensory
feedback
that
is
often
multimodal.
Based
on
deviations
between
predictions
and
actual
input,
primary
areas
of
cortex
have
been
shown
to
compute
sensorimotor
prediction
errors.
How
errors
one
modality
influence
the
computation
another
still
unclear.
To
investigate
multimodal
mouse
auditory
(ACx),
we
used
a
virtual
environment
experimentally
couple
running
both
self-generated
visual
feedback.
Using
two-photon
microscopy,
first
characterized
responses
layer
2/3
(L2/3)
neurons
sounds,
stimuli,
onsets
found
all
three
stimuli.
Probing
evoked
by
audiomotor
mismatches,
they
closely
resemble
visuomotor
mismatch
(V1).
Finally,
testing
for
cross
modal
coupling
sound
amplitude
flow
speed
running,
were
amplified
when
paired
with
concurrent
mismatches.
results
demonstrate
non-hierarchical
interactions
shape
error
cortical
L2/3.
***
Dear
reader,
please
note
this
manuscript
formatted
standard
submission
format,
statistical
information
provided
Table
S1.
Our
movements
result
in
predictable
sensory
feedback
that
is
often
multimodal.
Based
on
deviations
between
predictions
and
actual
input,
primary
areas
of
cortex
have
been
shown
to
compute
sensorimotor
prediction
errors.
How
errors
one
modality
influence
the
computation
another
still
unclear.
To
investigate
multimodal
mouse
auditory
cortex,
we
used
a
virtual
environment
experimentally
couple
running
both
self-generated
visual
feedback.
Using
two-photon
microscopy,
first
characterized
responses
layer
2/3
(L2/3)
neurons
sounds,
stimuli,
onsets
found
all
three
stimuli.
Probing
evoked
by
audiomotor
(AM)
mismatches,
they
closely
resemble
visuomotor
(VM)
mismatch
(V1).
Finally,
testing
for
cross
modal
AM
coupling
sound
amplitude
flow
speed
running,
were
amplified
when
paired
with
concurrent
VM
mismatches.
results
demonstrate
non-hierarchical
interactions
shape
error
cortical
L2/3.