Engineer friends often ask me: Graph Deep Learning sounds great, but are there any big commercial success stories? Is it being deployed in practical applications? Besides the obvious ones–recommendation systems at Pinterest, Alibaba and Twitter–a slightly nuanced success story is the Transformer architecture, which has taken the NLP industry by storm. Through this post, I want to establish links between Graph Neural Networks (GNNs) and Transformers. I’ll talk about the intuitions behind model architectures in the NLP and GNN communities, make connections using equations and figures, and discuss how we could work together to drive progress.
I. Golan, and R. El-Yaniv. Proceedings of the 32nd International Conference on Neural Information Processing Systems, page 9781–9791. Red Hook, NY, USA, Curran Associates Inc., (2018)
M. Ferrari Dacrema, P. Cremonesi, and D. Jannach. Proceedings of the 13th ACM Conference on Recommender Systems, page 101–109. New York, NY, USA, Association for Computing Machinery, (2019)
J. Weyn, D. Durran, and R. Caruana. Journal of Advances in Modeling Earth Systems, 11 (8):
2680--2693(2019)\_eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1029/2019MS001705.
N. Kimura, M. Kono, and J. Rekimoto. Proceedings of the 2019 CHI Conference on Human Factors in
Computing Systems, Paper 146, page 1--11. New York, NY, USA, Association for Computing Machinery, (May 2019)