AutoGNN: End-to-End Hardware-Driven Graph Preprocessing for Enhanced GNN Performance
Graph neural networks (GNNs) face significant bottlenecks in preprocessing, which often dominates overall inference latency. We introduce AutoGNN, an FPGA-based accelerator designed to address these challenges by leveraging FPGA’s reconfigurability and specialized components. AutoGNN adapts to diverse graph inputs, efficiently performing computationally intensive tasks such as graph conversion and sampling. By utilizing components like adder trees, AutoGNN executes reduction operations in a constant time, overcoming the limitations of serialization and synchronization on general-purpose GPUs. AutoGNN integrates unified processing elements (UPEs) and single-cycle reducers (SCRs) to streamline GNN preprocessing. UPEs enable scalable parallel processing for edge sorting and unique vertex selection, while SCRs efficiently handle sequential tasks such as pointer array construction and subgraph reindexing. A user-level software framework dynamically profiles graph inputs, determines optimal configurations, and reprograms AutoGNN to handle varying workloads. Implemented on a 7nm enterprise FPGA, AutoGNN achieves up to 9.0$\times$ and 2.1$\times$ speedup compared to conventional and GPU-accelerated preprocessing systems, respectively, enabling high-performance GNN preprocessing across diverse datasets.
Wed 4 FebDisplayed time zone: Hobart change
09:50 - 11:10 | Graph Neural Networks and Retrieval SystemsMain Conference at Coogee Chair(s): Amir Yazdanbakhsh Google Research, Brain Team | ||
09:50 20mTalk | VeloxGNN: Accelerating Out-of-Core based GNN Training with Low Data Migration and High Accuracy via Delayed Gradient Propagation Main Conference Yi Li University of Texas at Dallas, Tsun-Yu Yang Center for Computational Evolutionary Intelligence, Electrical & Computer Engineering, Duke University, Zhaoyan Shen Shandong University, Ming-Chang Yang The Chinese University of Hong Kong (CUHK), Bingzhe Li University of Texas at Dallas | ||
10:10 20mTalk | AutoGNN: End-to-End Hardware-Driven Graph Preprocessing for Enhanced GNN Performance Main Conference Seungkwan Kang KAIST, Seungjun Lee KAIST, Donghyun Gouk Panmnesia, Miryeong Kwon Panmnesia, Hyunkyu Choi Panmnesia, Junhyeok Jang Panmnesia, Sangwon Lee Panmnesia, Huiwon Choi KAIST, Jie Zhang Peking University, Wonil Choi Hanyang University, Mahmut Taylan Kandemir Pennsylvania State University, Myoungsoo Jung KAIST | ||
10:30 20mTalk | Scaling Graph Neural Network Training via Geometric Optimization Main Conference Fangzhou Ye University of Central Florida, Lingxiang Yin University of Central Florida, Hao Zheng University of Central Florida | ||
10:50 20mTalk | VectorLiteRAG: Latency-Aware and Fine-Grained Resource Partitioning for Efficient RAG Main Conference | ||