Learning in Spiking Neural Assemblies
–Neural Information Processing Systems
We consider a statistical framework for learning in a class of networks ofspiking neurons. Our aim is to show how optimal local learning rules can be readily derived once the neural dynamics and desired functionality of the neural assembly have been specified, in contrast to other models which assume (sub-optimal) learning rules. Within this framework we derive local rules for learning temporal sequencesin a model of spiking neurons and demonstrate its superior performance to correlation (Hebbian) based approaches. We further show how to include mechanisms such as synaptic depression andoutline how the framework is readily extensible to learning in networks of highly complex spiking neurons. A stochastic quantalvesicle release mechanism is considered and implications on the complexity of learning discussed.
Neural Information Processing Systems
Dec-31-2003