Loading [MathJax]/extensions/MathMenu.js
DARe: DropLayer-Aware Manycore ReRAM architecture for Training Graph Neural Networks | IEEE Conference Publication | IEEE Xplore

DARe: DropLayer-Aware Manycore ReRAM architecture for Training Graph Neural Networks


Abstract:

Graph Neural Networks (GNNs) are a variant of Deep Neural Networks (DNNs) operating on graphs. GNNs have attributes of both DNNs and graph computation. However, training ...Show More

Abstract:

Graph Neural Networks (GNNs) are a variant of Deep Neural Networks (DNNs) operating on graphs. GNNs have attributes of both DNNs and graph computation. However, training GNNs on manycore architectures is a challenging task because it involves heavy communication that bottlenecks performance. DropEdge and Dropout, which we collectively refer to as DropLayer, are regularization techniques that can improve the predictive accuracy of GNNs. Moreover, when implemented on a manycore architecture, DropEdge and Dropout are capable of reducing the on-chip traffic. In this paper, we present a ReRAM-based 3D manycore architecture called DARe, tailored for accelerating on-chip training of GNNs. The key component of the DARe architecture is a Network-on-Chip (NoC) that reduces the amount of communication using DropLayer. The reduced traffic prevents communication hotspots and leads to better performance. We demonstrate that DARe outperforms conventional GPUs by up to 6.7X (5.6X on average) in terms of execution time, while being up to 30X (23X on average) more energy efficient for GNN training.
Date of Conference: 01-04 November 2021
Date Added to IEEE Xplore: 23 December 2021
ISBN Information:

ISSN Information:

Conference Location: Munich, Germany

Funding Agency:

References is not available for this document.

I. Introduction

Graph Neural Networks (GNNs) are used for predictive analytics using graph-structured data. This makes them different from traditional Deep Neural Networks (DNNs) that operate on regular data structures such as images or sequences. GNN s have various real-life applications such as recommendation systems [1], quantum chemistry [2], social networks [3] [4] etc. To learn representation using the relational structure of graphs, GNN s perform iterative neighborhood aggregation, where each node aggregates features of its neighbors to compute new features [5]. This gives rise to repeated message-passing operations. GNN s exhibit characteristics of both DNN training (involving trainable weights) and graph analytics (accumulating neighboring vertices' information along graph edges). Hence, GNN training is both compute- and communication-intensive.

Select All
1.
R. Ying et al., "Graph Convolutional Neural Networks for Web-Scale Recommender Systems", ACM SIGKDD International Conference on Knowledge Discovery Data Mining, 2018.
2.
F. Ding, Graph Neural Networks for Quantum Chemistry, July 2019, [online] Available: https://github.com/ifding/graph-neural-networks.
3.
W. Fan et al., "Graph Neural Networks for Social Recommendation", The World Wide Web Conference, 2019.
4.
J. Zhou et al., "Graph Neural Networks: A Review of Methods", arXiv: 1812. 08434, 2018.
5.
M. Zhang et al., "An End-to-End Deep Learning Architecture for Graph Classification" in AAAI, New Orleans, LA, 2018.
6.
D. Fujiki, S. Mahlke and R. Das, "In-memory Data Flow Processor", International Conference on Parallel Architectures and Compilation Techniques (PACT), 2017.
7.
K. Kiningham, C. Re and P. Levis, GRIP: A Graph Neural Network Accelerator Architecture, 2020.
8.
Y. Rong, W. Huang, T. Xu and J. Huang, "DropEdge: Towards Deep Graph Convolutional Networks on Node Classification", International Conference on Learning Representations (ICLR), 2020.
9.
N. Srivastava et al., "Dropout: a simple way to prevent neural networks from overfitting", The Journal of Machine Learning Research, vol. 15, no. 1, pp. 1929-1958, 2014.
10.
T. N. Kipf and M. Welling, "Semi-Supervised Classification with Graph Convolutional Networks", International Conference on Learning Representations (ICLR), 2017.
11.
T. Geng et al., "AWB-GCN: A Graph Convolutional Network Accelerator with Runtime Workload Rebalancing", IEEE/ACM International Symposium on Microarchitecture (MICRO), 2020.
12.
B. Feero and P. P. Pande, "Networks-on-Chip in a Three-Dimensional Environment: A Performance Evaluation", IEEE Transactions on Computers, vol. 58, no. 1, pp. 32-45, 2009.
13.
W. L. Chiang et al., "Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks", ACM SIGKDD International Conference on Knowledge Discovery Data Mining, 2019.
14.
A. Auten, M. Tomei and R. Kumar, "Hardware Acceleration of Graph Neural Networks", IEEE/ACM Design Automation Conference (DAC), 2020.
15.
M. Yan et al., "HyGCN: A GCN Accelerator with Hybrid Architecture", IEEE International Symposium on High Performance Computer Architecture (HPCA), 2020.
16.
L. Song et al., "GraphR: Accelerating Graph Processing Using ReRAM", IEEE International Symposium on High-Performance Computer Architecture (HPCA), 2018.
17.
G. Dai et al., "GraphSAR: a sparsity-aware processing-in-memory architecture for large-scale graph processing on ReRAMs", Asia and South Pacific Design Automation Conference (ASP-DAC), 2019.
18.
L. Zheng et al., "Spara: An Energy-Efficient ReRAM-Based Accelerator for Sparse Graph Analytics Applications", IEEE International Parallel Distributed Processing Symposium (IPDPS), 2020.
19.
A. Shafiee et al., "ISAAC: a convolutional neural network accelerator with in-situ analog arithmetic in crossbars.", International Symposium on Computer Architecture (ISCA), 2016.
20.
L. Song et al., "PipeLayer: A Pipelined ReRAM-Based Accelerator for Deep Learning", IEEE International Symposium on High-Performance Computer Architecture (HPCA), 2017.
21.
P. Chi et al., "PRIME: A Novel Processing-in-Memory Architecture for Neural Network Computation in ReRAM-Based Main Memory", International Symposium on Computer Architecture (ISCA), 2016.
22.
Z. He, J. Lin, R. Ewetz, I. Yuan and D. Fan, "Noise Injection Adaption: End-to-End ReRAM Crossbar Non-ideal Effect Adaption for Neural Network Mapping", IEEE/ACM Design Automation Conference (DAC), 2019.
23.
B. K. Joardar et al., "AccuReD: High Accuracy Training of CNNs on ReRAM/GPU Heterogeneous 3-D Architecture", IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, vol. 40, no. 5, pp. 971-984, 2020.
24.
A. I. Arka et al., "ReGraphX: NoC-enabled 3D Heterogeneous ReRAM Architecture for Training Graph Neural Networks", Design Automation and Test in Europe Conference and Exhibition (DATE), 2021.
25.
Using variable length packets to embed extra network control information, 2010.
26.
J. Duato, S. Yalamanchil and N. Lionel, Interconnection Networks: An Engineering Approach, San Francisco, CA, USA:Morgan Kaufmann Publishers Inc, vol. 54, pp. 1025-1040, 2002.
27.
Y. Long, T. Na and S. Mukhopadhyay, "ReRAM-Based Processing-in-Memory Architecture for Recurrent Neural Network Acceleration", IEEE Transactions on Very Large Scale Integration (VLSI) Systems, vol. 26, no. 12, pp. 2781-2794, 2018.
28.
N. Agarwal, T. Krishna, L. Peh and N. Jha, "GARNET: A detailed on-chip network model inside a full-system simulator", Proc. of the ISP ASS, 2009.
29.
M. Bushnell and V. Agrawal, Essentials of Electronic Testing for Digital Memory and Mixed-Signal VLSI Circuits, Springer US, 2006.
30.
G. Karypis and V. Kumar, "A Fast and High Quality Multilevel Scheme for Partitioning Irregular Graphs", SIAM Journal on Scientific Computing, vol. 20, no. 6, pp. 359-392, 1998.
Contact IEEE to Subscribe

References

References is not available for this document.