ConceptNet Numberbatch consists of state-of-the-art semantic vectors (also known as word embeddings) that can be used directly as a representation of word meanings or as a starting point for further machine learning.
In natural language understanding, there is a hierarchy of lenses through which we can extract meaning - from words to sentences to paragraphs to documents. At the document level, one of the most useful ways to understand text is by analyzing its topics.
I made an introductory talk on word embeddings in the past and this write-up is an extended version of the part about philosophical ideas behind word vectors.
W. Zou, R. Socher, D. Cer, and C. Manning. Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, page 1393--1398. (2013)
G. Marco Baroni, Georgiana Dinu. 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014 - Proceedings of the Conference, (2014)
L. Hettinger, A. Zehe, A. Dallmann, and A. Hotho. INFORMATIK 2019: 50 Jahre Gesellschaft für Informatik – Informatik für Gesellschaft, page 191-204. Bonn, Gesellschaft für Informatik e.V., (2019)