Training and deploying GNNs to handle real-world graph data poses several theoretical and engineering challenges: 1. Giant Graphs – Memory Limitations 2. Sparse Computations – Hardware Limitations 3. Graph Subsampling – Reliability Limitations
The blogpost introduces three simple but effective ideas in the 🛠 'toolbox' 🛠 for developing efficient and scalable GNNs: 1. Graph data preparation and sampling 2. Efficient GNN architecture design 3. Learning Paradigms to improve performance and latency
The key idea: Sentences are fully-connected graphs of words, and Transformers are very similar to Graph Attention Networks (GATs) which use multi-head attention to aggregate features from their neighborhood nodes (i.e., words).
cc. @PetarV_93
This connection known to most people, but I've missed having all the information in one place. So I wrote it myself.