TY - UNPB
T1 - Efficient Synaptic Delay Implementation in Digital Event-Driven AI Accelerators
AU - Meijer, Roy
AU - Detterer, Paul
AU - Yousefzadeh, Amirreza
AU - Patino-Saucedo, Alberto
AU - Tang, Guanghzi
AU - Vadivel, Kanishkan
AU - Xu, Yinfu
AU - Gomony, Manil-Dev
AU - Corradi, Federico
AU - Linares-Barranco, Bernabe
AU - Sifalakis, Manolis
N1 - arXiv admin note: substantial text overlap with arXiv:2404.10597
PY - 2025/1/23
Y1 - 2025/1/23
N2 - Synaptic delay parameterization of neural network models have remained largely unexplored but recent literature has been showing promising results, suggesting the delay parameterized models are simpler, smaller, sparser, and thus more energy efficient than similar performing (e.g. task accuracy) non-delay parameterized ones. We introduce Shared Circular Delay Queue (SCDQ), a novel hardware structure for supporting synaptic delays on digital neuromorphic accelerators. Our analysis and hardware results show that it scales better in terms of memory, than current commonly used approaches, and is more amortizable to algorithm-hardware co-optimizations, where in fact, memory scaling is modulated by model sparsity and not merely network size. Next to memory we also report performance on latency area and energy per inference.
AB - Synaptic delay parameterization of neural network models have remained largely unexplored but recent literature has been showing promising results, suggesting the delay parameterized models are simpler, smaller, sparser, and thus more energy efficient than similar performing (e.g. task accuracy) non-delay parameterized ones. We introduce Shared Circular Delay Queue (SCDQ), a novel hardware structure for supporting synaptic delays on digital neuromorphic accelerators. Our analysis and hardware results show that it scales better in terms of memory, than current commonly used approaches, and is more amortizable to algorithm-hardware co-optimizations, where in fact, memory scaling is modulated by model sparsity and not merely network size. Next to memory we also report performance on latency area and energy per inference.
KW - cs.NE
KW - cs.AI
U2 - 10.48550/arXiv.2501.13610
DO - 10.48550/arXiv.2501.13610
M3 - Preprint
BT - Efficient Synaptic Delay Implementation in Digital Event-Driven AI Accelerators
PB - ArXiv.org
ER -