Abstract
Graph neural networks (GNNs) have emerged as a popular choice for analyzing structured data organized as graphs. Nevertheless, GNN models tend to be shallow, failing to fully exploit the capabilities of modern GPUs. Our motivational tests reveal that GPU dataloading for GNN inference yields remarkable performance enhancements when both the graph topology and features reside in GPU memory. Unfortunately, the use of this approach is hindered by the large size of real-world graph datasets. To address this limitation, we introduce GDL-GNN, a partition-based method that incorporates all essential information for inference within each subgraph. It thus combines the efficiency of GPU dataloading with layerwise inference, while maintaining the accuracy of full-neighbor inference. Additional optimization enables GDL-GNN to avoid unnecessary representation computation on halo nodes and to conceal file loading time. Evaluation shows the effectiveness of GDL-GNN in both single- and multi-GPU scenarios, revealing a reduction in inference time of up to 59.9% without compromising accuracy.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
Open-sourced at https://github.com/danghr/GDL-GNN.
References
Abadal, S., Jain, A., Guirado, R., López-Alonso, J., Alarcón, E.: Computing graph neural networks: a survey from algorithms to accelerators. ACM Comput. Surv. 54(9), 1–38 (2021)
Bai, Y., et al.: Efficient data loader for fast sampling-based GNN training on large graphs. IEEE Trans. Parallel Distrib. Syst. 32(10), 2541–2556 (2021)
DGL Team: DGL Documentation (2023). https://docs.dgl.ai/en/1.1.x/
Duan, K., et al.: A comprehensive study on large-scale graph training: benchmarking and rethinking. Adv. Neural Inf. Process. Syst. 35 (2022)
Fey, M., Lenssen, J.E.: Fast graph representation learning with PyTorch geometric. In: ICLR Workshop on Representation Learning on Graphs and Manifolds (2019)
Hamilton, W., Ying, Z., Leskovec, J.: Inductive representation learning on large graphs. Adv. Neural Inf. Process. Syst. 30 (2017)
Hu, W., et al.: Open graph benchmark: datasets for machine learning on graphs. Adv. Neural Inf. Process. Syst. 33 (2020)
Jangda, A., Polisetty, S., Guha, A., Serafini, M.: Accelerating graph sampling for graph machine learning using GPUs. In: Proceedings of the Sixteenth European Conference on Computer Systems (2021)
Juenger, D., Iskos, N., Wang, Y., Hemstad, J., Hundt, C., Sakharnykh, N.: Maximizing performance with massively parallel hash maps on GPUs (2023). https://developer.nvidia.com/blog/maximizing-performance-with-massively-parallel-hash-maps-on-gpus/
Karypis, G., Kumar, V.: A fast and high quality multilevel scheme for partitioning irregular graphs. SIAM J. Sci. Comput. 20(1), 359–392 (1998)
Kim, D.H., Nagi, R., Chen, D.: Thanos: high-performance CPU-GPU based balanced graph partitioning using cross-decomposition. In: 2020 25th Asia and South Pacific Design Automation Conference (2020)
Lin, H., et al.: A comprehensive survey on distributed training of graph neural networks. Proc. IEEE 111(12), 1572–1606 (2023)
Lin, Z., Li, C., Miao, Y., Liu, Y., Xu, Y.: PaGraph: scaling GNN training on large graphs via computation-aware caching. In: Proceedings of the 11th ACM Symposium on Cloud Computing (2020)
Liu, X., et al.: Survey on graph neural network acceleration: an algorithmic perspective. In: Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (2022)
Lv, Z., et al.: A survey of graph pre-processing methods: from algorithmic to hardware perspectives (2023). arXiv:2309.07581
Min, S.W., et al.: Large graph convolutional network training with GPU-oriented data communication architecture. Proc. VLDB Endow. 14(11) (2021)
PyG Team: PyG Documentation (2023). https://pytorch-geometric.readthedocs.io/en/2.4.0/
Song, S., Jiang, P.: Rethinking graph data placement for graph neural network training on multiple GPUs. In: Proceedings of the 36th ACM International Conference on Supercomputing (2022)
Veličković, P., Cucurull, G., Casanova, A., Romero, A., Liò, P., Bengio, Y.: Graph attention networks. In: 6th International Conference on Learning Representations (2018)
Wang, M., et al.: Deep graph library: a graph-centric, highly-performant package for graph neural networks (2020). arXiv:1909.01315
Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., Yu, P.S.: A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 32(1), 4–24 (2021)
Zhang, D., et al.: InferTurbo: a scalable system for boosting full-graph inference of graph neural network over huge graphs. In: 2023 IEEE 39th International Conference on Data Engineering (2023)
Zheng, D., et al.: DistDGL: distributed graph neural network training for billion-scale graphs. In: 2020 IEEE/ACM 10th Workshop on Irregular Applications: Architectures and Algorithms (2020)
Zhou, H., Srivastava, A., Zeng, H., Kannan, R., Prasanna, V.: Accelerating large scale real-time GNN inference using channel pruning. Proc. VLDB Endow. 14(9) (2021)
Zhou, J., et al.: Graph neural networks: a review of methods and applications. AI Open 1, 57–81 (2020)
Zhu, J., et al.: Simplifying distributed neural network training on massive graphs: randomized partitions improve model aggregation (2023). arXiv:2305.09887
Acknowledgement
The authors sincerely appreciates the anonymous reviewers for their valuable comments and suggestions. This work was supported by the National Natural Science Foundation of China (Grant No. 62202451), CAS Project for Young Scientists in Basic Research (Grant No. YSBR-029), and CAS Project for Youth Innovation Promotion Association.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Dang, H., Wu, M., Yan, M., Ye, X., Fan, D. (2024). GDL-GNN: Applying GPU Dataloading of Large Datasets for Graph Neural Network Inference. In: Carretero, J., Shende, S., Garcia-Blas, J., Brandic, I., Olcoz, K., Schreiber, M. (eds) Euro-Par 2024: Parallel Processing. Euro-Par 2024. Lecture Notes in Computer Science, vol 14802. Springer, Cham. https://doi.org/10.1007/978-3-031-69766-1_24
Download citation
DOI: https://doi.org/10.1007/978-3-031-69766-1_24
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-69765-4
Online ISBN: 978-3-031-69766-1
eBook Packages: Computer ScienceComputer Science (R0)