COIN: Communication-Aware In-Memory Acceleration for Graph Convolutional Networks

Sumit K. Mandal, Gokul Krishnan, A. Alper Goksoy, Gopikrishnan Ravindran Nair, Yu Cao, Umit Y. Ogras

Research output: Contribution to journalArticlepeer-review

8 Scopus citations

Abstract

Graph convolutional networks (GCNs) have shown remarkable learning capabilities when processing graph-structured data found inherently in many application areas. GCNs distribute the outputs of neural networks embedded in each vertex over multiple iterations to take advantage of the relations captured by the underlying graphs. Consequently, they incur a significant amount of computation and irregular communication overheads, which call for GCN-specific hardware accelerators. To this end, this paper presents a communication-aware in-memory computing architecture (COIN) for GCN hardware acceleration. Besides accelerating the computation using custom compute elements (CE) and in-memory computing, COIN aims at minimizing the intra- and inter-CE communication in GCN operations to optimize the performance and energy efficiency. Experimental evaluations with widely used datasets show up to 105times improvement in energy consumption compared to state-of-the-art GCN accelerator.

Original languageEnglish (US)
Pages (from-to)472-485
Number of pages14
JournalIEEE Journal on Emerging and Selected Topics in Circuits and Systems
Volume12
Issue number2
DOIs
StatePublished - Jun 1 2022
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2011 IEEE.

Keywords

  • Machine learning
  • graph neural networks
  • processing-in-memory
  • resistive RAM

Fingerprint

Dive into the research topics of 'COIN: Communication-Aware In-Memory Acceleration for Graph Convolutional Networks'. Together they form a unique fingerprint.

Cite this