CS224W: Machine Learning with Graphs | 2021 | Lecture 17.3 - Cluster GCN Scaling up GNNs

Jure Leskovec Computer Science, PhD Neighbor Sampling, presented in the previous lecture (17.2) constructs a computational graph separately for each node in a mini-batch. This creates a lot of redundancy in computing node embeddings within the mini-batch. A different approach is to sample a subgraph from a large graph that is small enough to be loaded into GPU. Then, the efficient and non-redundant full-batch GNN can be applied over the sampled subgraph. An example of this method is Cluster-GCN. Cluster-GCN first pre-processes a large graph by partitioning it into clusters of nodes. Then, during training, it samples clusters of nodes in each mini-batch and applies the full-batch GNN over the induced subgraph. To follow along with the course schedule and syllabus, visit: To get the latest news on Stanford’s upcoming professional programs in Artificial Intelligence, visit: To view all online courses and programs offered by Stanford,
Back to Top