I. Introduction
Recent advancements in Deep Neural Networks (DNNs) have demonstrated significant success across various applications. However, the increasing complexity and capabilities of DNNs necessitate substantial computational power and memory bandwidth in conventional Von Neumann architectures to accelerate DNN applications. A promising alternative lies in the utilization of novel architectures constructed with emerging technologies. Among the various options, the Resistive RAM (RRAM) crossbar-based architecture, comprised of memristor cells [1], emerges as an innovative compute-in-memory solution that not only reduces power consumption, but also boosts processing speeds. Illustrated in Fig. 1 is a standard RRAM crossbar. Within the crossbar, DNN kernels are unfolded and embedded with each memristor cells, each of which retaining a single weight value, while input data is continuously streamed into the crossbar from its wordlines. The analog nature of this architecture makes it well-suited for vector-matrix multiplication, as the dot product operation can be replicated using Kirchhoff’s circuit law.