Learning Working Memory in Recurrent Spiking Neural Networks Using Heterogeneous Delays

Laurent Perrinet

Cerco seminar

[2026-04-16]

logo

Contact me @ laurent.perrinet@univ-amu.fr

Polychronization

Polychronization

Polychronization

Polychronization

Methods : BPTT (snn Torch) - synthetic target

Methods : Weight initialization

$$ I_j(t) = \sum_{i=1}^{N} \bigl ( \sum_{d=1}^{D} \mathbf{W}_{j, i, d} \cdot s_i(t-d) \bigr ) $$ $$ u_j(t) = \beta \cdot u_j(t-1) \cdot (1 - s_j(t-1)) + I_j(t) $$ $$ s_j(t) = \mathbf{H}[u_j(t) \geq \vartheta] $$ $$ \mathbf{W} \mathbf{C} \approx \mathbf{S} $$ $$ w_{j, i, d} = \frac{1}{N \cdot D \cdot p_A \cdot M} \sum_{\mu=1}^{M} \sum_{t=D+1}^{T} s_{j}^{\mu}(t) \cdot s_i^{\mu}(t-d) $$

Results : recall of target

Results : role of parameters

Results : memory retrieval

Results : memory retrieval

Learning Working Memory in Recurrent Spiking Neural Networks Using Heterogeneous Delays

Laurent Perrinet

Cerco seminar

[2026-04-16]

logo

Contact me @ laurent.perrinet@univ-amu.fr