HPCA 2026
Sat 31 January - Wed 4 February 2026 Sydney, Australia
co-located with HPCA/CGO/PPoPP/CC 2026
Wed 4 Feb 2026 09:50 - 10:10 at Coogee - Graph Neural Networks and Retrieval Systems Chair(s): Amir Yazdanbakhsh

Training Graph Neural Networks (GNNs) on large-scale data is essential in domains such as transportation, recommendation systems, and molecular biology. As graph sizes increasingly surpass main memory capacities, the out-of-core (OOC) systems have been proposed, storing graphs on external storage, sequentially loading and processing smaller partitions. However, existing OOC solutions face key challenges, including excessive data migration between storage and memory, as well as reduced model accuracy. In this paper, we theoretically and empirically analyze the limitations of state-of-the-art (SOTA) OOC-based GNN systems and identify opportunities for optimization. Guided by our theoretical insights, we design a novel system, named VeloxGNN, to improve data migration efficiency while maintaining high model accuracy. First, we introduce a novel algorithm called Delayed Gradient Propagation (DGP), specifically designed for OOC-based GNN training. DGP leverages both historical node embeddings and gradients to achieve two key objectives: minimizing data migration (by ensuring that the graph is read at most once) while maintaining model accuracy. To support DGP, we then propose system-level optimizations, including dynamic memory management, a DGP-aware loading order, and a new graph partitioning scheme that separates labeled and unlabeled data. The experimental results in real systems show that VeloxGNN achieves in-memory model accuracy while reducing training time by 17.7% to 73.3% across various datasets and GNN models, outperforming state-of-the-art methods. This result highlights VeloxGNN’s potential for efficient and scalable GNN training on large-scale graph data.

Wed 4 Feb

Displayed time zone: Hobart change

09:50 - 11:10
Graph Neural Networks and Retrieval SystemsMain Conference at Coogee
Chair(s): Amir Yazdanbakhsh Google Research, Brain Team
09:50
20m
Talk
VeloxGNN: Accelerating Out-of-Core based GNN Training with Low Data Migration and High Accuracy via Delayed Gradient Propagation
Main Conference
Yi Li University of Texas at Dallas, Tsun-Yu Yang Center for Computational Evolutionary Intelligence, Electrical & Computer Engineering, Duke University, Zhaoyan Shen Shandong University, Ming-Chang Yang The Chinese University of Hong Kong (CUHK), Bingzhe Li University of Texas at Dallas
10:10
20m
Talk
AutoGNN: End-to-End Hardware-Driven Graph Preprocessing for Enhanced GNN Performance
Main Conference
Seungkwan Kang KAIST, Seungjun Lee KAIST, Donghyun Gouk Panmnesia, Miryeong Kwon Panmnesia, Hyunkyu Choi Panmnesia, Junhyeok Jang Panmnesia, Sangwon Lee Panmnesia, Huiwon Choi KAIST, Jie Zhang Peking University, Wonil Choi Hanyang University, Mahmut Taylan Kandemir Pennsylvania State University, Myoungsoo Jung KAIST
10:30
20m
Talk
Scaling Graph Neural Network Training via Geometric Optimization
Main Conference
Fangzhou Ye University of Central Florida, Lingxiang Yin University of Central Florida, Hao Zheng University of Central Florida
10:50
20m
Talk
VectorLiteRAG: Latency-Aware and Fine-Grained Resource Partitioning for Efficient RAG
Main Conference
Junkyum Kim Georgia Institute of Technology, Divya Mahajan Georgia Institute of Technology