Sparse Attentive Backtracking: Temporal Credit Assignment Through Reminding
Ke, Nan Rosemary, GOYAL, Anirudh Goyal ALIAS PARTH, Bilaniuk, Olexa, Binas, Jonathan, Mozer, Michael C., Pal, Chris, Bengio, Yoshua
–Neural Information Processing Systems
Learning long-term dependencies in extended temporal sequences requires credit assignment to events far back in the past. The most common method for training recurrent neural networks, back-propagation through time (BPTT), requires credit information to be propagated backwards through every single step of the forward computation, potentially over thousands or millions of time steps. This becomes computationally expensive or even infeasible when used with long sequences. Importantly, biological brains are unlikely to perform such detailed reverse replay over very long sequences of internal states (consider days, months, or years.) However, humans are often reminded of past memories or mental states which are associated with the current mental state.
Neural Information Processing Systems
Feb-14-2020, 19:56:34 GMT
- Technology: