@lepsky

A primer on word embedding

, , and . Data Intelligence and Cognitive Informatics, page 525--541. Singapore, Springer, (2021)
DOI: 10.1007/978-981-15-8530-2_42

Abstract

The current research on the topic of machine learning and especially the domain of natural language processing has gained much popularity in the modern era. One such framework for attaining NLP tasks is word embedding, which represents data as vectors, i.e., real numbers rather than words of natural language because neural networks do not understand them naturally. Word embeddings try to capture both syntactic and semantic information of words and capture relationships according to context and morphology. This paper reviews each word embedding technique available in the contemporary world ranging from traditional embeddings based on the frequency of terms to pre-trained embeddings like prediction-based embeddings. The goal of this paper is to present the myriad methods available for word embedding, classify their working patterns, also identify their pros and cons for working on text classification and detect their hegemony over the traditional methods of NLP.

Links and resources

Tags