VeloxGNN: Accelerating Out-of-Core based GNN Training with Low Data Migration and High Accuracy via Delayed Gradient Propagation
Training Graph Neural Networks (GNNs) on large-scale data is essential in domains such as transportation, recommendation systems, and molecular biology. As graph sizes increasingly surpass main memory capacities, the out-of-core (OOC) systems have been proposed, storing graphs on external storage, sequentially loading and processing smaller partitions. However, existing OOC solutions face key challenges, including excessive data migration between storage and memory, as well as reduced model accuracy. In this paper, we theoretically and empirically analyze the limitations of state-of-the-art (SOTA) OOC-based GNN systems and identify opportunities for optimization. Guided by our theoretical insights, we design a novel system, named VeloxGNN, to improve data migration efficiency while maintaining high model accuracy. First, we introduce a novel algorithm called Delayed Gradient Propagation (DGP), specifically designed for OOC-based GNN training. DGP leverages both historical node embeddings and gradients to achieve two key objectives: minimizing data migration (by ensuring that the graph is read at most once) while maintaining model accuracy. To support DGP, we then propose system-level optimizations, including dynamic memory management, a DGP-aware loading order, and a new graph partitioning scheme that separates labeled and unlabeled data. The experimental results in real systems show that VeloxGNN achieves in-memory model accuracy while reducing training time by 17.7% to 73.3% across various datasets and GNN models, outperforming state-of-the-art methods. This result highlights VeloxGNN’s potential for efficient and scalable GNN training on large-scale graph data.
Wed 4 FebDisplayed time zone: Hobart change
09:50 - 11:10 | Graph Neural Networks and Retrieval SystemsMain Conference at Coogee Chair(s): Amir Yazdanbakhsh Google Research, Brain Team | ||
09:50 20mTalk | VeloxGNN: Accelerating Out-of-Core based GNN Training with Low Data Migration and High Accuracy via Delayed Gradient Propagation Main Conference Yi Li University of Texas at Dallas, Tsun-Yu Yang Center for Computational Evolutionary Intelligence, Electrical & Computer Engineering, Duke University, Zhaoyan Shen Shandong University, Ming-Chang Yang The Chinese University of Hong Kong (CUHK), Bingzhe Li University of Texas at Dallas | ||
10:10 20mTalk | AutoGNN: End-to-End Hardware-Driven Graph Preprocessing for Enhanced GNN Performance Main Conference Seungkwan Kang KAIST, Seungjun Lee KAIST, Donghyun Gouk Panmnesia, Miryeong Kwon Panmnesia, Hyunkyu Choi Panmnesia, Junhyeok Jang Panmnesia, Sangwon Lee Panmnesia, Huiwon Choi KAIST, Jie Zhang Peking University, Wonil Choi Hanyang University, Mahmut Taylan Kandemir Pennsylvania State University, Myoungsoo Jung KAIST | ||
10:30 20mTalk | Scaling Graph Neural Network Training via Geometric Optimization Main Conference Fangzhou Ye University of Central Florida, Lingxiang Yin University of Central Florida, Hao Zheng University of Central Florida | ||
10:50 20mTalk | VectorLiteRAG: Latency-Aware and Fine-Grained Resource Partitioning for Efficient RAG Main Conference | ||