Comprehensive Design Space Exploration for Graph Neural Network Aggregation on GPUs

Hyunwoo Nam, Jay Hwan Lee, Shinhyung Yang, Yeonsoo Kim, Jiun Jeong, Jeonggeun Kim, Bernd Burgstaller

Research output: Contribution to journalArticlepeer-review

Abstract

Graph neural networks (GNNs) have become the state-of-the-art technology for extracting and predicting data representations on graphs. With increasing demand to accelerate GNN computations, the GPU has become the dominant platform for GNN training and inference. GNNs consist of a compute-bound combination phase and a memory-bound aggregation phase. The memory access patterns of the aggregation phase remain a major performance bottleneck on GPUs, despite recent microarchitectural enhancements. Although GNN characterizations have been conducted to investigate this bottleneck, they did not reveal the impact of architectural modifications. However, a comprehensive understanding of improvements from such modifications is imperative to devise GPU optimizations for the aggregation phase. In this letter, we explore the GPU design space for aggregation by assessing the performance improvement potential of a series of architectural modifications. We find that the low locality of aggregation deteriorates performance with increased thread-level parallelism, and a significant enhancement follows memory access optimizations, which remain effective even with software optimization. Our analysis provides insights for hardware optimizations to significantly improve GNN aggregation on GPUs.

Original languageEnglish
Pages (from-to)45-48
Number of pages4
JournalIEEE Computer Architecture Letters
Volume24
Issue number1
DOIs
Publication statusPublished - 2025

Bibliographical note

Publisher Copyright:
© 2025 IEEE.

All Science Journal Classification (ASJC) codes

  • Hardware and Architecture

Fingerprint

Dive into the research topics of 'Comprehensive Design Space Exploration for Graph Neural Network Aggregation on GPUs'. Together they form a unique fingerprint.

Cite this