Goto

Collaborating Authors

 weight update


Neuronal Competition Groups with Supervised STDP for Spike-Based Classification

Neural Information Processing Systems

Spike Timing-Dependent Plasticity (STDP) is a promising substitute to backprop-agation for local training of Spiking Neural Networks (SNNs) on neuromorphic hardware. STDP allows SNNs to address classification tasks by combining unsupervised STDP for feature extraction and supervised STDP for classification. Unsupervised STDP is usually employed with Winner-Takes-All (WT A) competition to learn distinct patterns.


Communication-efficientDistributedSGDwith Sketching

Neural Information Processing Systems

However,theoretical and empirical evidence both suggest that there is a maximum mini-batch size beyond which the number of iterations required toconvergestops decreasing, andgeneralization error begins toincrease [Maetal.,2017,Lietal., 2014, Golmant et al., 2018, Shallue et al., 2018, Keskar et al., 2016, Hoffer et al., 2017]. In this paper, we aim instead to decrease the communication cost per worker.






A Distinguishing supervised learning from reinforcement learning in a feedforward model { 1, 1} and t = 1,, T, are projected onto a hiddenlayer h

Neural Information Processing Systems

In order to illustrate the main idea from our paper in a simplified context, we show in this section how observed hidden-layer activity in a linear feedforward network can be used to infer the learning rule that is used to train the network. Consider the simple feedforward network shown in Fig. S1. N (0, Σ) is noise injected into the network. This is similar to learning with Feedback Alignment [4], except that here we do not assume that the readout weights are being learned. Equations (11) and (13) provide predictions for how the hidden-layer activity is expected to evolve under either SL or RL.