Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Accelerating GNNs on GPU Sparse Tensor Cores through N:M Sparsity-Oriented Graph Reordering

Conference ·
Recent advancements in GPU hardware support have introduced the capability to leverage N:M sparse patterns for substantial performance gains. Graphs in Graph Neural Networks (GNNs) are typically sparse, but the sparsity is often irregular, not conforming to such sparse patterns. In this paper, we propose a novel graph reordering algorithm, the first of its kind, to reshape irregular graph data into the N:M structured sparse pattern at the tile level, allowing linear-algebra-based graph operations in GNNs to benefit from the N:M sparse hardware. The optimization is lossless, maintaining the accuracy of GNN. It can remove 98-100\% violations of the N:M sparse patterns at the vector level, and increase the proportion of conforming graphs in SuiteSparse collection from 5-9\% to 88.7-93.5\%. On A100 GPUs, the optimization accelerates Sparse Matrix Matrix (SpMM) by up to 43X (2.3X -- 7.5X on average) and speeds up the key graph operations in GNNs on real graphs by as much as 8.6X (3.5X on average).
Research Organization:
Pacific Northwest National Laboratory (PNNL), Richland, WA (United States)
Sponsoring Organization:
USDOE
DOE Contract Number:
AC05-76RL01830
OSTI ID:
2545648
Report Number(s):
PNNL-SA-193030
Country of Publication:
United States
Language:
English

Similar Records

Accelerating GNNs on GPU Sparse Tensor Cores through N:M Sparsity-Oriented Graph Reordering
Conference · Thu Feb 27 23:00:00 EST 2025 · OSTI ID:2524569

Adapting In Situ Accelerators for Sparsity With Granular Matrix Reordering
Journal Article · Tue Oct 20 20:00:00 EDT 2020 · IEEE Computer Architecture Letters · OSTI ID:1691454

pnnl/emp-gnn
Software · Wed Feb 28 19:00:00 EST 2024 · OSTI ID:code-123162