MULTIMODAL MISMATCH RESPONSES IN MOUSE AUDITORY CORTEX DOI Creative Commons
Magdalena Sołyga, Georg B. Keller

bioRxiv (Cold Spring Harbor Laboratory), Год журнала: 2023, Номер unknown

Опубликована: Ноя. 1, 2023

SUMMARY Our movements result in predictable sensory feedback that is often multimodal. Based on deviations between predictions and actual input, primary areas of cortex have been shown to compute sensorimotor prediction errors. How errors one modality influence the computation another still unclear. To investigate multimodal mouse auditory (ACx), we used a virtual environment experimentally couple running both self-generated visual feedback. Using two-photon microscopy, first characterized responses layer 2/3 (L2/3) neurons sounds, stimuli, onsets found all three stimuli. Probing evoked by audiomotor mismatches, they closely resemble visuomotor mismatch (V1). Finally, testing for cross modal coupling sound amplitude flow speed running, were amplified when paired with concurrent mismatches. results demonstrate non-hierarchical interactions shape error cortical L2/3.

Язык: Английский

Prefrontal Cortex subregions provide distinct visual and behavioral feedback modulation to the Primary Visual Cortex DOI Creative Commons
Sofie Ährlund‐Richter, Yuma Osako, Kyle Jenks

и другие.

bioRxiv (Cold Spring Harbor Laboratory), Год журнала: 2024, Номер unknown

Опубликована: Авг. 8, 2024

Abstract The mammalian Prefrontal Cortex (PFC) has been suggested to modulate sensory information processing across multiple cortical regions via long-range axonal projections. These projections arise from PFC subregions with unique brain-wide connectivity and functional repertoires, which may provide the architecture for modular feedback intended shape processing. Here, we used tracing, somatic 2-photon calcium imaging, chemogenetic manipulations in mice delineate how Anterior Cingulate (ACA) ventrolateral Orbitofrontal (ORB) of primary Visual (VISp) behavioral states. Structurally, found that ACA ORB have distinct patterning both layers. axons VISp had a stronger representation visual stimulus than axons, but showed non-visual, behavior-dependent activity. input enhanced encoding stimuli by neurons, modulation responses scaled arousal. On other hand, shaped movement arousal related responses, specifically reduced high-contrast stimuli. Thus, separable projection patterns encode information, putatively providing substrate their effects on representations VISp. Our results offer refined model hierarchy its impact processing, whereby as opposed generalized properties contribute activity during discrete

Язык: Английский

Процитировано

1

Cortical acetylcholine dynamics are predicted by cholinergic axon activity and behavior state DOI Creative Commons
Erin Neyhart, Na Zhou, Brandon R. Munn

и другие.

bioRxiv (Cold Spring Harbor Laboratory), Год журнала: 2023, Номер unknown

Опубликована: Ноя. 16, 2023

Abstract Even under spontaneous conditions and in the absence of changing environmental demands, awake animals alternate between increased or decreased periods alertness. These changes brain state can occur rapidly, on a timescale seconds, neuromodulators such as acetylcholine (ACh) are thought to play an important role driving these transitions. Here, we perform first simultaneous imaging ACh sensors GCaMP-expressing axons vivo , examine spatiotemporal properties cortical activity release during behavioral state. We observed high correlation simultaneously recorded basal forebrain axon neuromodulator sensor fluorescence around locomotion pupil dilation. Consistent with volume transmission ACh, increases were accompanied by local levels that fell off distance from nearest axon. GRAB-ACh could be accurately predicted axonal alone, providing validation is reliable proxy for nearby levels. Deconvolution traces allowed us account kinetics emphasized rapid clearance smaller transients outside running periods. Finally, trained predictive model fluctuations combination size speed; this performed better than using either variable generalized well unseen data. Overall, results contribute growing understanding precise timing spatial characteristics fast

Язык: Английский

Процитировано

2

Quantification of the effect of hemodynamic occlusion in two-photon imaging DOI Creative Commons

Baba Yogesh,

Matthias Heindorf, Rebecca Jordan

и другие.

bioRxiv (Cold Spring Harbor Laboratory), Год журнала: 2024, Номер unknown

Опубликована: Окт. 29, 2024

ABSTRACT The last few years have seen an explosion in the number of tools available to measure neuronal activity using fluorescence imaging (Chen et al., 2013; Feng 2019; Jing Sun 2018; Wan 2021). When performed vivo, these measurements are invariably contaminated by hemodynamic occlusion artifacts. In widefield calcium imaging, this problem is well recognized. For two-photon however, effects only been sparsely characterized. Here we perform a quantification changes observed with GFP expression both and imaging. We find that many instances magnitude signal attributable comparable sensors. Moreover, were spatially heterogeneous, over cortical regions across depth, exhibited complex relationship behavior. Thus, important caveat consider when analyzing interpreting not just but also data.

Язык: Английский

Процитировано

0

MULTIMODAL MISMATCH RESPONSES IN MOUSE AUDITORY CORTEX DOI Creative Commons
Magdalena Sołyga, Georg B. Keller

bioRxiv (Cold Spring Harbor Laboratory), Год журнала: 2023, Номер unknown

Опубликована: Ноя. 1, 2023

SUMMARY Our movements result in predictable sensory feedback that is often multimodal. Based on deviations between predictions and actual input, primary areas of cortex have been shown to compute sensorimotor prediction errors. How errors one modality influence the computation another still unclear. To investigate multimodal mouse auditory (ACx), we used a virtual environment experimentally couple running both self-generated visual feedback. Using two-photon microscopy, first characterized responses layer 2/3 (L2/3) neurons sounds, stimuli, onsets found all three stimuli. Probing evoked by audiomotor mismatches, they closely resemble visuomotor mismatch (V1). Finally, testing for cross modal coupling sound amplitude flow speed running, were amplified when paired with concurrent mismatches. results demonstrate non-hierarchical interactions shape error cortical L2/3.

Язык: Английский

Процитировано

0