Skip to main content

GDL-GNN: Applying GPU Dataloading of Large Datasets for Graph Neural Network Inference

  • Conference paper
  • First Online:
Euro-Par 2024: Parallel Processing (Euro-Par 2024)

Abstract

Graph neural networks (GNNs) have emerged as a popular choice for analyzing structured data organized as graphs. Nevertheless, GNN models tend to be shallow, failing to fully exploit the capabilities of modern GPUs. Our motivational tests reveal that GPU dataloading for GNN inference yields remarkable performance enhancements when both the graph topology and features reside in GPU memory. Unfortunately, the use of this approach is hindered by the large size of real-world graph datasets. To address this limitation, we introduce GDL-GNN, a partition-based method that incorporates all essential information for inference within each subgraph. It thus combines the efficiency of GPU dataloading with layerwise inference, while maintaining the accuracy of full-neighbor inference. Additional optimization enables GDL-GNN to avoid unnecessary representation computation on halo nodes and to conceal file loading time. Evaluation shows the effectiveness of GDL-GNN in both single- and multi-GPU scenarios, revealing a reduction in inference time of up to 59.9% without compromising accuracy.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    Open-sourced at https://github.com/danghr/GDL-GNN.

References

  1. Abadal, S., Jain, A., Guirado, R., López-Alonso, J., Alarcón, E.: Computing graph neural networks: a survey from algorithms to accelerators. ACM Comput. Surv. 54(9), 1–38 (2021)

    Article  Google Scholar 

  2. Bai, Y., et al.: Efficient data loader for fast sampling-based GNN training on large graphs. IEEE Trans. Parallel Distrib. Syst. 32(10), 2541–2556 (2021)

    Article  Google Scholar 

  3. DGL Team: DGL Documentation (2023). https://docs.dgl.ai/en/1.1.x/

  4. Duan, K., et al.: A comprehensive study on large-scale graph training: benchmarking and rethinking. Adv. Neural Inf. Process. Syst. 35 (2022)

    Google Scholar 

  5. Fey, M., Lenssen, J.E.: Fast graph representation learning with PyTorch geometric. In: ICLR Workshop on Representation Learning on Graphs and Manifolds (2019)

    Google Scholar 

  6. Hamilton, W., Ying, Z., Leskovec, J.: Inductive representation learning on large graphs. Adv. Neural Inf. Process. Syst. 30 (2017)

    Google Scholar 

  7. Hu, W., et al.: Open graph benchmark: datasets for machine learning on graphs. Adv. Neural Inf. Process. Syst. 33 (2020)

    Google Scholar 

  8. Jangda, A., Polisetty, S., Guha, A., Serafini, M.: Accelerating graph sampling for graph machine learning using GPUs. In: Proceedings of the Sixteenth European Conference on Computer Systems (2021)

    Google Scholar 

  9. Juenger, D., Iskos, N., Wang, Y., Hemstad, J., Hundt, C., Sakharnykh, N.: Maximizing performance with massively parallel hash maps on GPUs (2023). https://developer.nvidia.com/blog/maximizing-performance-with-massively-parallel-hash-maps-on-gpus/

  10. Karypis, G., Kumar, V.: A fast and high quality multilevel scheme for partitioning irregular graphs. SIAM J. Sci. Comput. 20(1), 359–392 (1998)

    Article  MathSciNet  Google Scholar 

  11. Kim, D.H., Nagi, R., Chen, D.: Thanos: high-performance CPU-GPU based balanced graph partitioning using cross-decomposition. In: 2020 25th Asia and South Pacific Design Automation Conference (2020)

    Google Scholar 

  12. Lin, H., et al.: A comprehensive survey on distributed training of graph neural networks. Proc. IEEE 111(12), 1572–1606 (2023)

    Article  Google Scholar 

  13. Lin, Z., Li, C., Miao, Y., Liu, Y., Xu, Y.: PaGraph: scaling GNN training on large graphs via computation-aware caching. In: Proceedings of the 11th ACM Symposium on Cloud Computing (2020)

    Google Scholar 

  14. Liu, X., et al.: Survey on graph neural network acceleration: an algorithmic perspective. In: Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (2022)

    Google Scholar 

  15. Lv, Z., et al.: A survey of graph pre-processing methods: from algorithmic to hardware perspectives (2023). arXiv:2309.07581

  16. Min, S.W., et al.: Large graph convolutional network training with GPU-oriented data communication architecture. Proc. VLDB Endow. 14(11) (2021)

    Google Scholar 

  17. PyG Team: PyG Documentation (2023). https://pytorch-geometric.readthedocs.io/en/2.4.0/

  18. Song, S., Jiang, P.: Rethinking graph data placement for graph neural network training on multiple GPUs. In: Proceedings of the 36th ACM International Conference on Supercomputing (2022)

    Google Scholar 

  19. Veličković, P., Cucurull, G., Casanova, A., Romero, A., Liò, P., Bengio, Y.: Graph attention networks. In: 6th International Conference on Learning Representations (2018)

    Google Scholar 

  20. Wang, M., et al.: Deep graph library: a graph-centric, highly-performant package for graph neural networks (2020). arXiv:1909.01315

  21. Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., Yu, P.S.: A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 32(1), 4–24 (2021)

    Article  MathSciNet  Google Scholar 

  22. Zhang, D., et al.: InferTurbo: a scalable system for boosting full-graph inference of graph neural network over huge graphs. In: 2023 IEEE 39th International Conference on Data Engineering (2023)

    Google Scholar 

  23. Zheng, D., et al.: DistDGL: distributed graph neural network training for billion-scale graphs. In: 2020 IEEE/ACM 10th Workshop on Irregular Applications: Architectures and Algorithms (2020)

    Google Scholar 

  24. Zhou, H., Srivastava, A., Zeng, H., Kannan, R., Prasanna, V.: Accelerating large scale real-time GNN inference using channel pruning. Proc. VLDB Endow. 14(9) (2021)

    Google Scholar 

  25. Zhou, J., et al.: Graph neural networks: a review of methods and applications. AI Open 1, 57–81 (2020)

    Article  Google Scholar 

  26. Zhu, J., et al.: Simplifying distributed neural network training on massive graphs: randomized partitions improve model aggregation (2023). arXiv:2305.09887

Download references

Acknowledgement

The authors sincerely appreciates the anonymous reviewers for their valuable comments and suggestions. This work was supported by the National Natural Science Foundation of China (Grant No. 62202451), CAS Project for Young Scientists in Basic Research (Grant No. YSBR-029), and CAS Project for Youth Innovation Promotion Association.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mingyu Yan .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Dang, H., Wu, M., Yan, M., Ye, X., Fan, D. (2024). GDL-GNN: Applying GPU Dataloading of Large Datasets for Graph Neural Network Inference. In: Carretero, J., Shende, S., Garcia-Blas, J., Brandic, I., Olcoz, K., Schreiber, M. (eds) Euro-Par 2024: Parallel Processing. Euro-Par 2024. Lecture Notes in Computer Science, vol 14802. Springer, Cham. https://doi.org/10.1007/978-3-031-69766-1_24

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-69766-1_24

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-69765-4

  • Online ISBN: 978-3-031-69766-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics