World Wide relies on analytics signals to operate securely and keep research services available. Accept to continue, or leave the site.
Review the Privacy Policy for details about analytics processing.
Dr
Princeton
Showing your local timezone
Schedule
Wednesday, May 3, 2023
5:00 PM Europe/Berlin
Seminar location
No geocoded details are available for this content yet.
Meeting Password
$Em4HF
Use this password when joining the live session
Format
Past Seminar
Recording
Not available
Host
SNUFA
Duration
30.00 minutes
Seminar location
No geocoded details are available for this content yet.
Neural activity is often described in terms of population-level factors extracted from the responses of many neurons. Factors provide a lower-dimensional description with the aim of shedding light on network computations. Yet, mechanistically, computations are performed not by continuously valued factors but by interactions among neurons that spike discretely and variably. Models provide a means of bridging these levels of description. We developed a general method for training model networks of spiking neurons by leveraging factors extracted from either data or firing-rate-based networks. In addition to providing a useful model-building framework, this formalism illustrates how reliable and continuously valued factors can arise from seemingly stochastic spiking. Our framework establishes procedures for embedding this property in network models with different levels of realism. The relationship between spikes and factors in such networks provides a foundation for interpreting (and subtly redefining) commonly used quantities such as firing rates.
Brian DePasquale
Dr
Princeton
neuro
Decades of research on understanding the mechanisms of attentional selection have focused on identifying the units (representations) on which attention operates in order to guide prioritized sensory p
neuro
neuro