Relational data represent relationships between entities anywhere on the web (e.g. online social networks) or in the physical world (e.g. structure of the protein).
Graph neural networks are intimately related to partial differential equations governing information diffusion on graphs. Thinking of GNNs as PDEs leads to a new broad class of graph ML methods.
TL;DR: Have you even wondered what is so special about convolution? In this post, I derive the convolution from first principles and show that it naturally emerges from translational symmetry. During…
Engineer friends often ask me: Graph Deep Learning sounds great, but are there any big commercial success stories? Is it being deployed in practical applications? Besides the obvious ones–recommendation systems at Pinterest, Alibaba and Twitter–a slightly nuanced success story is the Transformer architecture, which has taken the NLP industry by storm. Through this post, I want to establish links between Graph Neural Networks (GNNs) and Transformers. I’ll talk about the intuitions behind model architectures in the NLP and GNN communities, make connections using equations and figures, and discuss how we could work together to drive progress.
Fullstack GraphQL Tutorial to go from zero to production covering all basics and advanced concepts. Includes tutorials for Apollo, Relay, React and NodeJS.
A. Bosselut, H. Rashkin, M. Sap, C. Malaviya, A. Celikyilmaz, and Y. Choi. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, page 4762--4779. Florence, Italy, Association for Computational Linguistics, (July 2019)
W. Hamilton, R. Ying, and J. Leskovec. (2017)cite arxiv:1709.05584Comment: Published in the IEEE Data Engineering Bulletin, September 2017; version with minor corrections.
J. Qiu, Y. Dong, H. Ma, J. Li, K. Wang, and J. Tang. Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining - WSDM \textquotesingle18, ACM Press, (2018)
D. Gibson, J. Kleinberg, and P. Raghavan. Proceedings of the ninth ACM conference on Hypertext and hypermedia : links, objects, time and space---structure in hypermedia systems links, objects, time and space---structure in hypermedia systems - HYPERTEXT \textquotesingle98, ACM Press, (1998)
A. Bosselut, H. Rashkin, M. Sap, C. Malaviya, A. Celikyilmaz, and Y. Choi. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, page 4762--4779. Florence, Italy, Association for Computational Linguistics, (July 2019)
N. Du, M. Wang, L. Tran, G. Lee, and I. Shafran. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), page 4980--4991. Hong Kong, China, Association for Computational Linguistics, (November 2019)
Z. Zhang, X. Han, Z. Liu, X. Jiang, M. Sun, and Q. Liu. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, page 1441--1451. Florence, Italy, Association for Computational Linguistics, (July 2019)
R. Kumar, A. Tomkins, and E. Vee. WSDM '08: Proceedings of the international conference on Web search and web data mining, page 129--138. New York, NY, USA, ACM, (2008)
G. Gossen, E. Demidova, and T. Risse. Proceedings of the 15th ACM/IEEE-CS Joint Conference on Digital Libraries, page 75--84. New York, NY, USA, ACM, (2015)
J. Lee, R. Rossi, and X. Kong. Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, page 1666--1674. ACM, (2018)