Research Salon

Accurate online training of dynamical spiking neural networks through Forward Propagation Through Time

https://arxiv.org/abs/2112.11231

Background

  • Training Recurrent Neural Networks via Forward Propagation Through Time
    https://icml.cc/virtual/2021/spotlight/10384
     
  • Liquid Time-constant Networks https://arxiv.org/abs/2006.04439
     
  • SNNs have demonstrated competitive and energy-efficient applications so far, e.g. Yin et al. 2021

Backprop Through Time (BPTT)

  • Training instabilities: vanishing or exploding gradients

     
  • Memory overhead Ω(T)
  • cannot be applied online: network is 'locked' during backward pass

Forward Propagation Through Time (FPTT)

  • FPTT takes a gradient step to minimize an instantaneous risk function at each time step

  • straightforward application of FPTT to SNNs fails

FPTT objective functions

Intermediate Losses for Terminal Prediction

with

Pseudo code

Liquid Time Constant SNN

Hasani et al., 2020

--> time constants are a function of inputs and hidden states

The Add task

One data point consists of two sequences (x1, x2) of length T and a target label y.

Example:

x1: <0,    0,    0,    1,     0,    0,    1,    0>

x2: <0.1, 0.6, 0.7, 0.2, 0.5, 0.8, 0.4, 0.3>

y:   <0,     0,    0,    0.2, 0.2, 0.2, 0.6, 0.6>

ASRNN: Adaptive Spiking RNN

LTC-SRNN: Liquid Time Constant Spiking RNN

The Add task

DVS Gesture

  • 20 - 500 frames are pre-processed using an SCNN
  • pre-processed input is flattened to 1D inputs

FPTT improves over online approximations of BPTT

Memory requirements and training time

Conclusions

  • FPTT works very well for long sequence learning with recurrent SNNs as long as LTC neurons are used.
     
  • It's not slower (with room for improvement) and has constant memory requirements.
     
  • Potentially very interesting in an online setting on neuromorphic hardware

Open Questions

  • How stable is training really?
     
  • Could this be applied to non-recurrent architectures?
     
  • Could this be combined with self-supervised learning?

FPTT research salon

By Gregor Lenz

FPTT research salon

  • 159