Large-Scale Graph Neural Networks:
Navigating the Past and Pioneering New Horizons
AAAI 2024
AAAI 2024
This tutorial represents a notable achievement as it offers a comprehensive overview of techniques designed for large-scale machine learning on graphs, encompassing both theoretical foundations and practical applications. It delves into past and recent research endeavors aimed at enhancing the scalability of Graph Neural Networks (GNNs) and explores their diverse potential use cases. This tutorial caters to a broad audience, targeting engineers, researchers, graduate students, and industry professionals keen on harnessing scalable GNNs for large-scale datasets. After the tutorial, the audience is expected to learn both the foundational theory of classical and novel model frameworks, as well as their applications.
Abstract
Graph Neural Networks (GNNs) have gained significant attention in recent years due to their ability to model complex relationships between entities in graph-structured data such as social networks, protein structures, and knowledge graphs. However, due to the size of real-world industrial graphs and the special architecture of GNNs, it is a long-lasting challenge for engineers and researchers to deploy GNNs on large-scale graphs, which significantly limits their applications in real-world applications. In this tutorial, we will cover the fundamental scalability challenges of GNNs, frontiers of large-scale GNNs including classic approaches and some newly emerging techniques, the evaluation and comparison of scalable GNNs, and their large-scale real-world applications. Overall, this tutorial aims to provide a systematic and comprehensive understanding of the challenges and state-of-the-art techniques for scaling GNNs. The summary and discussion on future directions will inspire engineers and researchers to explore new ideas and developments in this rapidly evolving field.
Outline
1. Introduction of GNNs (30 minutes)
(a) Foundations of GNNs
(b) Applications of GNNs
(c) Scalability Challenges of Large-Scale GNNs
2. Classic Approaches for Scaling GNNs (60 minutes)
(a) Sampling Methods
(b) Decoupling Methods
(c) Distributed Methods
Break: 30 minutes
3. Emerging Techniques for Scaling GNNs (60 minutes)
Training:
(a) Lazy Propagation
(b) Alternating Training
(c) GNN Pre-training
Inferencing:
Cross-model Distillation
Data:
(a) Graph Condensation
(b) Subgraph Sketching
(c) Tabularization
4. Evaluation, Comparison and Applications (20 minutes)
5. Summary and Future Directions (10 minutes)
Presentation Details
We first offer a brief introduction to graph neural networks (GNNs). We will illustrate several classical and basic GNN models and their real-world applications [1–3]. These domains include social networks [4], biological molecules [5], and recommendation systems [6]. Following the basic concepts, we will introduce the scalability challenge of large-scale GNNs via both theoretical analysis and empirical examples [7, 8].
This section focuses on classic research on the scalability and efficiency of large-scale GNNs using various innovative designs, such as sampling methods [7–9], pre/post-computing methods [3,10,11], and distributed methods [12] . Targeting on different categories, we will introduce their basic methodologies, how they mitigate the scalability issue, their variants, and their imitations. This part provides the audience with the necessaryknowledge about how we can work with graphs of increasing sizes.
3. Emerging Techniques for Scaling GNNs
This part will provide an overview of emerging trends and techniques in scalable GNN research. We will discuss the latest developments in three perspectives: training, inferencing and data. In terms of training strategies, we delve into innovative approaches such as lazy graph propagation [13], alternating training [14], and the growing importance of pre-training [15]. In the context of inference, we introduce the concept of Cross-model distillation [16]. Additionally, within the realm of data management, we explore evolving
techniques such as Graph condensation [17], Subgraph Sketching [18], and Tabularization [19]. This comprehensive examination provides valuable insights into the cutting-edge developments that are shaping the future of scalable GNN research. By the end of this discussion, the audience will gain a comprehensive understanding of the current state-of-the-art in scalable GNN research and their potential for solving real-world problems.
4. Evaluation, Comparison, and Applications
In this section, we will evaluate the schemes introduced, focusing on their accuracy and computational complexity. The outcome of these evaluations will inform readers about the advantages and drawbacks of each method. Such insights will not only guide the selection of suitable techniques for scaling GNNs but also inspire researchers and practitioners to create more effective and precise large-scale GNNs. As the availability of large-scale graph data grows, the importance of scalable GNNs for addressing intricate problems and making accurate forecasts in various fields will rise. We will delve into the implications of scalable GNNs in notable real-world applications, such as web-scale recommendation systems [20], friend
recommendation [21], and fraud detection. Additionally, we will introduce several platforms and packages tailored for large-scale graphs. To conclude, we will discuss some prevailing challenges and potential avenues for research in real-world applications.
5. Summary and Future Directions
We will summarize the tutorial and provide an discussion on the future directions of large-scale GNNs.
Neil Shah is a Lead Research Scientist and Manager at Snap Re- search, working on machine learning algorithms and applications on large-scale graph data. His work has resulted in 55+ conference and journal publications, in top venues such as ICLR, NeurIPS, KDD, WSDM, WWW, AAAI and more, including several best-paper awards. He has also served as an organizer, chair and senior pro- gram committee member at a number of these conferences. He has also organized workshops and tutorials on graph machine learning topics at KDD, WSDM, SDM, ICDM, CIKM, and WWW. He has had previous research experiences at Lawrence Livermore National Laboratory, Microsoft Research, and Twitch. He earned a PhD in Computer Science in 2017 from Carnegie Mellon University’s Computer Science Department, funded partially by the NSF Graduate Research Fellowship.
Tutorial Slides
REFERENCES
Kipf, T. N., & Welling, M. (2016). Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907.
Gasteiger, J., Bojchevski, A., & Günnemann, S. (2018). Predict then propagate: Graph neural networks meet personalized PageRank. arXiv preprint arXiv:1810.05997.
Wu, F., Souza, A., Zhang, T., Fifty, C., Yu, T., & Weinberger, K. (2019). Simplifying graph convolutional networks. In International conference on machine learning, pages 6861–6871. PMLR.
Yao, L., Mao, C., & Luo, Y. (2019). Graph convolutional networks for text classification. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pages 7370–7377.
Zhang, Y., Zhang, H., & Tang, J. (2018). Graph convolutional networks for protein structure prediction. In Advances in Neural Information Processing Systems, pages 991–1001.
Wu, S., Sun, F., Zhang, W., Xie, X., & Cui, B. (2022). Graph neural networks in recommender systems: a survey. ACM Computing Surveys, 55(5), 1–37.
Hamilton, W., Ying, Z., & Leskovec, J. (2017). Inductive representation learning on large graphs. In Advances in neural information processing systems, 30.
Chen, J., Ma, T., & Xiao, C. (2018). Fastgcn: fast learning with graph convolutional networks via importance sampling. arXiv preprint arXiv:1801.10247.
Chiang, W. L., Liu, X., Si, S., Li, Y., Bengio, S., & Hsieh, C. J. (2019). Cluster-GCN: An efficient algorithm for training deep and large graph convolutional networks. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pages 257–266.
Sun, C., Gu, H., & Hu, J. (2021). Scalable and adaptive graph neural networks with self-label-enhanced training. arXiv preprint arXiv:2104.09376.
Huang, Q., He, H., Singh, A., Lim, S. N., & Benson, A. R. (2020). Combining label propagation and simple models out-performs graph neural networks. arXiv preprint arXiv:2010.13993.
Vasimuddin, Md., Misra, S., Ma, G., Mohanty, R., Georganas, E., Heinecke, A., Kalamkar, D., Ahmed, N. K., & Avancha, S. (2021). DistGNN: Scalable distributed training for large-scale graph neural networks. In Proceedings of the International Conference for High-Performance Computing, Networking, Storage and Analysis, pages 1–14.
Xue, R., Han, H., Torkamani, M. A., Pei, J., & Liu, X. (2023). LazyGNN: Large-scale graph neural networks via lazy propagation. ICML, 2023.
Han, H., Liu, X., Torkamani, A., Shi, F., Lee, V., & Tang, J. (2023). Alternately optimized graph neural networks. ICML, 2023.
Han, X., Zhao, T., Liu, Y., Hu, X., & Shah, N. (2023). MLPInit: Embarrassingly simple GNN training acceleration with MLP initialization. In The Eleventh International Conference on Learning Representations, 2023.
Zhang, S., Liu, Y., Sun, Y., & Shah, N. (2021). Graph-less neural networks: Teaching old MLPs new tricks via distillation. arXiv preprint arXiv:2110.08727.
Jin, W., Zhao, L., Zhang, S., Liu, Y., Tang, J., & Shah, N. (2021). Graph condensation for graph neural networks. arXiv preprint arXiv:2110.07580. 2021
Chamberlain, B. P., Shirobokov, S., Rossi, E., Frasca, F., Markovich, T., Hammerla, N., Bronstein, M. M., & Hansmire, M. (2022). Graph neural networks for link prediction with subgraph sketching. arXiv preprint arXiv:2209.15486.
Zhang, D., Huang, X., Liu, Z., Hu, Z., Song, X., Ge, Z., Zhang, Z., Wang, L., Zhou, J., Shuang, Y., et al. (2020). AGL: A scalable system for industrial-purpose graph machine learning. arXiv preprint arXiv:2003.02454.
Ying, R., He, R., Chen, K., Eksombatchai, P., Hamilton, W. L., & Leskovec, J. (2018). Graph convolutional neural networks for web-scale recommender systems. In Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, pages 974–983.
Shi, J., Chaurasiya, V., Liu, Y., Vij, S., Wu, Y., Kanduri, S., Shah, N., Yu, P., Srivastava, N., Shi, L., et al. (2023). Embedding based retrieval in friend recommendation. 2023