PLoS ONE,
Год журнала:
2025,
Номер
20(1), С. e0308848 - e0308848
Опубликована: Янв. 16, 2025
Immersive
virtual
reality
(VR)
environments
are
a
powerful
tool
to
explore
cognitive
processes
ranging
from
memory
and
navigation
visual
processing
decision
making-and
do
so
in
naturalistic
yet
controlled
setting.
As
such,
they
have
been
employed
across
different
species,
by
diverse
range
of
research
groups.
Unfortunately,
designing
implementing
behavioral
tasks
such
often
proves
complicated.
To
tackle
this
challenge,
we
created
DomeVR,
an
immersive
VR
environment
built
using
Unreal
Engine
4
(UE4).
UE4
is
game
engine
supporting
photo-realistic
graphics
containing
scripting
language
designed
for
use
non-programmers.
result,
easily
drag-and-drop
elements.
DomeVR
aims
make
these
features
accessible
neuroscience
experiments.
This
includes
logging
synchronization
system
solve
timing
uncertainties
inherent
UE4;
interactive
GUI
scientists
observe
subjects
during
experiments
adjust
task
parameters
on
the
fly,
dome
projection
full
immersion
non-human
subjects.
These
key
modular
can
be
added
individually
into
other
projects.
Finally,
present
proof-of-principle
data
highlighting
functionality
three
species:
human,
macaque
mouse.
Science,
Год журнала:
2025,
Номер
387(6730), С. 214 - 220
Опубликована: Янв. 9, 2025
The
current
understanding
of
primate
natural
action
organization
derives
from
laboratory
experiments
in
restrained
contexts
(RCs)
under
the
assumption
that
this
knowledge
generalizes
to
freely
moving
(FMCs).
In
work,
we
developed
a
neurobehavioral
platform
enable
wireless
recording
same
premotor
neurons
both
RCs
and
FMCs.
Neurons
often
encoded
hand
mouth
actions
differently
Furthermore,
FMCs,
identified
cells
selectively
untestable
during
others
displayed
mixed
selectivity
for
multiple
actions,
which
is
compatible
with
an
based
on
cortical
motor
synergies
at
different
levels
complexity.
Cross-context
decoding
demonstrated
neural
activity
FMCs
richer
more
generalizable
than
RCs,
suggests
neuroethological
approaches
are
better
suited
unveil
bases
behavior.
Nature Neuroscience,
Год журнала:
2023,
Номер
26(11), С. 1953 - 1959
Опубликована: Окт. 12, 2023
Organisms
process
sensory
information
in
the
context
of
their
own
moving
bodies,
an
idea
referred
to
as
embodiment.
This
is
important
for
developmental
neuroscience,
robotics
and
systems
neuroscience.
The
mechanisms
supporting
embodiment
are
unknown,
but
a
manifestation
could
be
observation
mice
brain-wide
neuromodulation,
including
primary
visual
cortex,
driven
by
task-irrelevant
spontaneous
body
movements.
We
tested
this
hypothesis
macaque
monkeys
(Macaca
mulatta),
primate
model
human
vision,
simultaneously
recording
cortex
activity
facial
also
sought
direct
comparison
using
analogous
approach
those
used
mouse
studies.
Here
we
found
that
(V1,
V2
V3/V3A)
was
associated
with
animals'
movements,
modulation
largely
explained
impact
movements
on
retinal
image,
is,
changes
input.
These
results
indicate
primates
minimally
may
reflect
species-specific
sensorimotor
strategies.
Current Opinion in Neurobiology,
Год журнала:
2024,
Номер
86, С. 102859 - 102859
Опубликована: Апрель 6, 2024
One
of
the
most
exciting
new
developments
in
systems
neuroscience
is
progress
being
made
toward
neurophysiological
experiments
that
move
beyond
simplified
laboratory
settings
and
address
richness
natural
behavior.
This
enabled
by
technological
advances
such
as
wireless
recording
freely
moving
animals,
automated
quantification
behavior,
methods
for
analyzing
large
data
sets.
Beyond
empirical
data,
however,
there
also
a
need
theories
concepts
to
interpret
data.
Such
particular
challenges
which
often
differ
significantly
from
scenarios
studied
traditional
settings.
Here,
we
discuss
some
strategies
developing
novel
example
hypotheses
proposed.
Nature Neuroscience,
Год журнала:
2024,
Номер
27(4), С. 772 - 781
Опубликована: Март 5, 2024
Abstract
Until
now,
it
has
been
difficult
to
examine
the
neural
bases
of
foraging
in
naturalistic
environments
because
previous
approaches
have
relied
on
restrained
animals
performing
trial-based
tasks.
Here
we
allowed
unrestrained
monkeys
freely
interact
with
concurrent
reward
options
while
wirelessly
recorded
population
activity
dorsolateral
prefrontal
cortex.
The
decided
when
and
where
forage
based
whether
their
prediction
was
fulfilled
or
violated.
This
not
solely
a
history
delivery,
but
also
understanding
that
waiting
longer
improves
chance
reward.
task
variables
were
continuously
represented
subspace
high-dimensional
activity,
this
compressed
representation
predicted
animal’s
subsequent
choices
better
than
true
as
well
raw
activity.
Our
results
indicate
monkeys’
strategies
are
cortical
model
dynamics
explore
environment.
Proceedings of the National Academy of Sciences,
Год журнала:
2025,
Номер
122(6)
Опубликована: Фев. 3, 2025
Our
understanding
of
how
vision
functions
as
primates
actively
navigate
the
real-world
is
remarkably
sparse.
As
most
data
have
been
limited
to
chaired
and
typically
head-restrained
animals,
synergistic
interactions
different
motor
actions/plans
inherent
active
sensing—e.g.,
eyes,
head,
posture,
movement,
etc.—on
visual
perception
are
largely
unknown.
To
address
this
considerable
gap
in
knowledge,
we
developed
an
innovative
wireless
head-mounted
eye-tracking
system
that
performs
Chair-free
Eye-Recording
using
Backpack
mounted
micROcontrollers
(CEREBRO)
for
small
mammals,
such
marmoset
monkeys.
Because
eye
illumination
environment
lighting
change
continuously
natural
contexts,
a
segmentation
artificial
neural
network
perform
robust
pupil
tracking
these
conditions.
Leveraging
investigate
vision,
demonstrate
although
freely
moving
marmosets
exhibit
frequent
compensatory
movements
equivalent
other
primates,
including
humans,
predictability
behavior
(gaze)
higher
when
animals
relative
they
head-fixed.
Moreover,
despite
increases
eye/head-motion
during
locomotion,
gaze
stabilization
remains
steady
because
increase
vestibularocular
reflex
gain
locomotion.
These
results
efficient,
dynamic
visuo-motor
mechanisms
related
behaviors
enable
stable,
high-resolution
foveal
explore
world.