Efficient Temporal Coding in the Early Visual System: Existing Evidence and Future Directions DOI Creative Commons

Byron H. Price,

Jeffrey P. Gavornik

Frontiers in Computational Neuroscience, Journal Year: 2022, Volume and Issue: 16

Published: July 4, 2022

While it is universally accepted that the brain makes predictions, there little agreement about how this accomplished and under which conditions. Accurate prediction requires neural circuits to learn store spatiotemporal patterns observed in natural environment, but not obvious such information should be stored, or encoded. Information theory provides a mathematical formalism can used measure efficiency utility of different coding schemes for data transfer storage. This shows codes become efficient when they remove predictable, redundant spatial temporal information. Efficient has been understand retinal computations may also relevant understanding more complicated processing visual cortex. However, literature on cortex varied confusing since same terms are mean things experimental theoretical contexts. In work, we attempt provide clear summary relationship between prediction, review evidence principles explain retina. We then apply framework occurring early visuocortical areas, arguing from rodents largely consistent with predictions model. Finally, respond criticisms suggest ways might design future experiments, particular focus extent make representations environmental statistics.

Language: Английский

Sound elicits stereotyped facial movements that provide a sensitive index of hearing abilities in mice DOI
Kameron K. Clayton, Kamryn S. Stecyk,

Anna Guo

et al.

Current Biology, Journal Year: 2024, Volume and Issue: 34(8), P. 1605 - 1620.e5

Published: March 15, 2024

Language: Английский

Citations

19

Eye saccades align optic flow with retinal specializations during object pursuit in freely moving ferrets DOI Creative Commons
Damian J. Wallace,

Kay-Michael Voit,

Danuza de Oliveira Machado

et al.

Current Biology, Journal Year: 2025, Volume and Issue: unknown

Published: Feb. 1, 2025

Highlights•Saccades during target pursuit align area centralis with intended direction of travel•Saccades simultaneously also retinal pattern optic flow•Post-saccade eye and head rotation reduce image blur limit information loss•Tree shrews, mice, rats have the same coordinated kineticsSummaryDuring prey pursuit, how rotations, such as saccades, enable continuous tracking erratically moving targets while enabling an animal to navigate through environment is unknown. To better understand this, we measured rotations in freely running ferrets behavior. By all environmental features, reconstructed animal's visual fields their relationship structures. In fields, position clustered on around high-acuity location, centralis, surprisingly, this cluster was not significantly shifted by digital removal either exclusively elicited when made turns, or that were tightly synchronized saccades. Here, show that, saccades did fixate they instead aligned travel. This features flow pattern, focus expansion, used for navigation many species. While initially rotated eyes turn, followed countering ongoing rotation, which reduced limited loss across field turns. As rotational tree rats, suggest these counter-rotations are a generalized mechanism mammals complex environments pursuit.Graphical abstract

Language: Английский

Citations

2

Active vision in freely moving marmosets using head-mounted eye tracking DOI Creative Commons
Vikram Pal Singh,

Jingwen Li,

Kyle Dawson

et al.

Proceedings of the National Academy of Sciences, Journal Year: 2025, Volume and Issue: 122(6)

Published: Feb. 3, 2025

Our understanding of how vision functions as primates actively navigate the real-world is remarkably sparse. As most data have been limited to chaired and typically head-restrained animals, synergistic interactions different motor actions/plans inherent active sensing—e.g., eyes, head, posture, movement, etc.—on visual perception are largely unknown. To address this considerable gap in knowledge, we developed an innovative wireless head-mounted eye-tracking system that performs Chair-free Eye-Recording using Backpack mounted micROcontrollers (CEREBRO) for small mammals, such marmoset monkeys. Because eye illumination environment lighting change continuously natural contexts, a segmentation artificial neural network perform robust pupil tracking these conditions. Leveraging investigate vision, demonstrate although freely moving marmosets exhibit frequent compensatory movements equivalent other primates, including humans, predictability behavior (gaze) higher when animals relative they head-fixed. Moreover, despite increases eye/head-motion during locomotion, gaze stabilization remains steady because increase vestibularocular reflex gain locomotion. These results efficient, dynamic visuo-motor mechanisms related behaviors enable stable, high-resolution foveal explore world.

Language: Английский

Citations

1

Detailed characterization of neural selectivity in free viewing primates DOI Creative Commons
Jacob L. Yates, Shanna Coop, Gabriel Sarch

et al.

Nature Communications, Journal Year: 2023, Volume and Issue: 14(1)

Published: June 20, 2023

Abstract Fixation constraints in visual tasks are ubiquitous and cognitive neuroscience. Despite its widespread use, fixation requires trained subjects, is limited by the accuracy of fixational eye movements, ignores role movements shaping input. To overcome these limitations, we developed a suite hardware software tools to study vision during natural behavior untrained subjects. We measured receptive fields tuning properties from multiple cortical areas marmoset monkeys who freely viewed full-field noise stimuli. The resulting curves primary cortex (V1) area MT match reported selectivity literature which was using conventional approaches. then combined free viewing with high-resolution tracking make first detailed 2D spatiotemporal measurements foveal V1. These findings demonstrate power characterize neural responses animals while simultaneously studying dynamics behavior.

Language: Английский

Citations

20

Interactions between rodent visual and spatial systems during navigation DOI
Aman B. Saleem, Laura Busse

Nature reviews. Neuroscience, Journal Year: 2023, Volume and Issue: 24(8), P. 487 - 501

Published: June 28, 2023

Language: Английский

Citations

20

A dynamic sequence of visual processing initiated by gaze shifts DOI
Philip R. L. Parker, Dylan M. Martins, Emmalyn S. P. Leonard

et al.

Nature Neuroscience, Journal Year: 2023, Volume and Issue: 26(12), P. 2192 - 2202

Published: Nov. 23, 2023

Language: Английский

Citations

20

Fast prediction in marmoset reach-to-grasp movements for dynamic prey DOI Creative Commons
Luke Shaw, Kuan Hong Wang, Jude F. Mitchell

et al.

Current Biology, Journal Year: 2023, Volume and Issue: 33(12), P. 2557 - 2565.e4

Published: June 1, 2023

Language: Английский

Citations

16

Coding of latent variables in sensory, parietal, and frontal cortices during closed-loop virtual navigation DOI Creative Commons
Jean‐Paul Noel, Edoardo Balzani, Eric Avila

et al.

eLife, Journal Year: 2022, Volume and Issue: 11

Published: Oct. 25, 2022

We do not understand how neural nodes operate and coordinate within the recurrent action-perception loops that characterize naturalistic self-environment interactions. Here, we record single-unit spiking activity local field potentials (LFPs) simultaneously from dorsomedial superior temporal area (MSTd), parietal 7a, dorsolateral prefrontal cortex (dlPFC) as monkeys navigate in virtual reality to ‘catch fireflies’. This task requires animals actively sample a closed-loop environment while concurrently computing continuous latent variables: (i) distance angle travelled (i.e., path integration) (ii) memorized firefly location hidden spatial goal). observed patterned mixed selectivity, with most prominently coding for variables, sensorimotor MSTd often eye movements. However, even traditionally considered sensory MSTd) tracked demonstrating integration vector of goals. Further, global encoding profiles unit-to-unit coupling noise correlations) suggested functional subnetwork composed by dlPFC, between these anatomy would suggest. show greater more animals’ gaze position was indicative ongoing goal. suggest this MSTd-dlPFC reflects monkeys’ natural adaptive strategy wherein they continuously toward (invisible) target. Together, results highlight distributed nature during closed fine-grain subnetworks may be dynamically established subserve (embodied) strategies.

Language: Английский

Citations

21

How ‘visual’ is the visual cortex? The interactions between the visual cortex and other sensory, motivational and motor systems as enabling factors for visual perception DOI
Cyriel M. A. Pennartz, Matthijs N. Oude Lohuis, Umberto Olcese

et al.

Philosophical Transactions of the Royal Society B Biological Sciences, Journal Year: 2023, Volume and Issue: 378(1886)

Published: Aug. 7, 2023

The definition of the visual cortex is primarily based on evidence that lesions this area impair perception. However, does not exclude may process more information than retinal origin alone, or other brain structures contribute to vision. Indeed, research across past decades has shown non-visual information, such as neural activity related reward expectation and value, locomotion, working memory sensory modalities, can modulate primary cortical responses inputs. Nevertheless, function poorly understood. Here we review recent evidence, coming from studies in rodents, arguing motor effects play a role processing itself, for instance disentangling direct auditory sound-evoked orofacial movement. These findings are placed broader framework casting vision terms predictive under control frontal, reward- motor-related systems. In contrast prevalent notion exclusively constructed by system, propose percepts generated larger network-the extended system-spanning cortices, supramodal areas frontal This article part theme issue 'Decision processes multisensory perception'.

Language: Английский

Citations

12

Cortical Integration of Vestibular and Visual Cues for Navigation, Visual Processing, and Perception DOI Creative Commons

Sepiedeh Keshavarzi,

Mateo Vélez‐Fort, Troy W. Margrie

et al.

Annual Review of Neuroscience, Journal Year: 2023, Volume and Issue: 46(1), P. 301 - 320

Published: July 10, 2023

Despite increasing evidence of its involvement in several key functions the cerebral cortex, vestibular sense rarely enters our consciousness. Indeed, extent to which these internal signals are incorporated within cortical sensory representation and how they might be relied upon for sensory-driven decision-making, during, example, spatial navigation, is yet understood. Recent novel experimental approaches rodents have probed both physiological behavioral significance indicate that their widespread integration with vision improves perceptual accuracy self-motion orientation. Here, we summarize recent findings a focus on circuits involved visual perception navigation highlight major remaining knowledge gaps. We suggest vestibulo-visual reflects process constant updating regarding status self-motion, access such information by cortex used predictions may implemented rapid, navigation-related decision-making.

Language: Английский

Citations

11