Goto

Collaborating Authors

 recurrent circuit


A Recurrent Neural Circuit Mechanism of T emporal-scaling Equivariant Representation

Neural Information Processing Systems

Time perception is fundamental in our daily life. An important feature of time perception is temporal scaling (TS): the ability to generate temporal sequences (e.g., movements) with different speeds. However, it is largely unknown about the mathematical principle underlying TS in the brain.


The Bayesian sampling in a canonical recurrent circuit with a diversity of inhibitory interneurons

Neural Information Processing Systems

Accumulating evidence suggests stochastic cortical circuits can perform sampling-based Bayesian inference to compute the latent stimulus posterior. Canonical cortical circuits consist of excitatory (E) neurons and types of inhibitory (I) in-terneurons. Nevertheless, nearly no sampling neural circuit models consider the diversity of interneurons, and thus how interneurons contribute to sampling remains poorly understood.


A Recurrent Neural Circuit Mechanism of T emporal-scaling Equivariant Representation

Neural Information Processing Systems

Time perception is fundamental in our daily life. An important feature of time perception is temporal scaling (TS): the ability to generate temporal sequences (e.g., movements) with different speeds. However, it is largely unknown about the mathematical principle underlying TS in the brain.


Information Dynamics and Emergent Computation in Recurrent Circuits of Spiking Neurons

Neural Information Processing Systems

We employ an efficient method using Bayesian and linear classifiers for analyzing the dynamics of information in high-dimensional states of generic cortical microcircuit models. It is shown that such recurrent cir- cuits of spiking neurons have an inherent capability to carry out rapid computations on complex spike patterns, merging information contained in the order of spike arrival with previously acquired context information.


Neural pathway crucial to successful rapid object recognition in primates

#artificialintelligence

MIT researchers have identified a brain pathway critical in enabling primates to effortlessly identify objects in their field of vision. The findings enrich existing models of the neural circuitry involved in visual perception and help to further unravel the computational code for solving object recognition in the primate brain. Led by Kohitij Kar, a postdoc at the McGovern Institute for Brain Research and Department of Brain and Cognitive Sciences, the study looked at an area called the ventrolateral prefrontal cortex (vlPFC), which sends feedback signals to the inferior temporal (IT) cortex via a network of neurons. The main goal of this study was to test how the back-and-forth information processing of this circuitry -- that is, this recurrent neural network -- is essential to rapid object identification in primates. The current study, published in Neuron and available via open access, is a followup to prior work published by Kar and James DiCarlo, the Peter de Florez Professor of Neuroscience, the head of MIT's Department of Brain and Cognitive Sciences, and an investigator in the McGovern Institute and the Center for Brains, Minds, and Machines.


Information Dynamics and Emergent Computation in Recurrent Circuits of Spiking Neurons

Natschläger, Thomas, Maass, Wolfgang

Neural Information Processing Systems

We employ an efficient method using Bayesian and linear classifiers for analyzing the dynamics of information in high-dimensional states of generic cortical microcircuit models. It is shown that such recurrent circuits of spiking neurons have an inherent capability to carry out rapid computations on complex spike patterns, merging information contained in the order of spike arrival with previously acquired context information.


Information Dynamics and Emergent Computation in Recurrent Circuits of Spiking Neurons

Natschläger, Thomas, Maass, Wolfgang

Neural Information Processing Systems

We employ an efficient method using Bayesian and linear classifiers for analyzing the dynamics of information in high-dimensional states of generic cortical microcircuit models. It is shown that such recurrent circuits of spiking neurons have an inherent capability to carry out rapid computations on complex spike patterns, merging information contained in the order of spike arrival with previously acquired context information.


Information Dynamics and Emergent Computation in Recurrent Circuits of Spiking Neurons

Natschläger, Thomas, Maass, Wolfgang

Neural Information Processing Systems

We employ an efficient method using Bayesian and linear classifiers for analyzing the dynamics of information in high-dimensional states of generic cortical microcircuit models. It is shown that such recurrent circuits of spiking neurons have an inherent capability to carry out rapid computations on complex spike patterns, merging information contained in the order of spike arrival with previously acquired context information.


A Model for Real-Time Computation in Generic Neural Microcircuits

Maass, Wolfgang, Natschläger, Thomas, Markram, Henry

Neural Information Processing Systems

A key challenge for neural modeling is to explain how a continuous stream of multi-modal input from a rapidly changing environment can be processed by stereotypical recurrent circuits of integrate-and-fire neurons in real-time. We propose a new computational model that is based on principles of high dimensional dynamical systems in combination with statistical learning theory. It can be implemented on generic evolved or found recurrent circuitry.


A Model for Real-Time Computation in Generic Neural Microcircuits

Maass, Wolfgang, Natschläger, Thomas, Markram, Henry

Neural Information Processing Systems

A key challenge for neural modeling is to explain how a continuous stream of multi-modal input from a rapidly changing environment can be processed by stereotypical recurrent circuits of integrate-and-fire neurons in real-time. We propose a new computational model that is based on principles of high dimensional dynamical systems in combination with statistical learning theory. It can be implemented on generic evolved or found recurrent circuitry.