I. Introduction
In the last decade, spiking neural networks (SNNs) have gained significant attention in the context of energy-efficient machine intelligence [1]. SNNs encode input data information with discrete binary spikes over multiple time-steps making them highly suitable for asynchronous event-driven input processing applications [2], [3]. Recent works have proposed full-scale general-purpose von-Neumann architectures leveraging the temporal processing property of SNNs [4], [5]. Other works, such as [6] and [7] have proposed novel dataflow to minimize the hardware overhead in von-Neumann implementation of SNNs. However, SNNs like conventional artificial neural networks (ANNs) entail significant dot-product operations leading to high memory and energy overhead when implemented on traditional von-Neumann architectures (due to the “memory wall bottleneck”) [8], [9]. To this end, analog in-memory computing (IMC) architectures [10], [11], [12] have been proposed to perform analog dot-product or multiply-and-accumulate (MAC) operations to achieve high memory bandwidth and compute parallelism, thereby overcoming the “memory wall bottleneck.”