The current research on the topic of machine learning and especially the domain of natural language processing has gained much popularity in the modern era. One such framework for attaining NLP tasks is word embedding, which represents data as vectors, i.e., real numbers rather than words of natural language because neural networks do not understand them naturally. Word embeddings try to capture both syntactic and semantic information of words and capture relationships according to context and morphology. This paper reviews each word embedding technique available in the contemporary world ranging from traditional embeddings based on the frequency of terms to pre-trained embeddings like prediction-based embeddings. The goal of this paper is to present the myriad methods available for word embedding, classify their working patterns, also identify their pros and cons for working on text classification and detect their hegemony over the traditional methods of NLP.
%0 Conference Paper
%1 satvika_primer_2021
%A Satvika,
%A Thada, Vikas
%A Singh, Jaswinder
%B Data Intelligence and Cognitive Informatics
%C Singapore
%D 2021
%E Jeena Jacob, I.
%E Kolandapalayam Shanmugam, Selvanayaki
%E Piramuthu, Selwyn
%E Falkowski-Gilski, Przemyslaw
%I Springer
%K maschinelles_lernen texttechnologie
%P 525--541
%R 10.1007/978-981-15-8530-2_42
%T A primer on word embedding
%X The current research on the topic of machine learning and especially the domain of natural language processing has gained much popularity in the modern era. One such framework for attaining NLP tasks is word embedding, which represents data as vectors, i.e., real numbers rather than words of natural language because neural networks do not understand them naturally. Word embeddings try to capture both syntactic and semantic information of words and capture relationships according to context and morphology. This paper reviews each word embedding technique available in the contemporary world ranging from traditional embeddings based on the frequency of terms to pre-trained embeddings like prediction-based embeddings. The goal of this paper is to present the myriad methods available for word embedding, classify their working patterns, also identify their pros and cons for working on text classification and detect their hegemony over the traditional methods of NLP.
%@ 9789811585302
@inproceedings{satvika_primer_2021,
abstract = {The current research on the topic of machine learning and especially the domain of natural language processing has gained much popularity in the modern era. One such framework for attaining NLP tasks is word embedding, which represents data as vectors, i.e., real numbers rather than words of natural language because neural networks do not understand them naturally. Word embeddings try to capture both syntactic and semantic information of words and capture relationships according to context and morphology. This paper reviews each word embedding technique available in the contemporary world ranging from traditional embeddings based on the frequency of terms to pre-trained embeddings like prediction-based embeddings. The goal of this paper is to present the myriad methods available for word embedding, classify their working patterns, also identify their pros and cons for working on text classification and detect their hegemony over the traditional methods of NLP.},
added-at = {2021-02-08T17:41:11.000+0100},
address = {Singapore},
author = {{Satvika} and Thada, Vikas and Singh, Jaswinder},
biburl = {https://www.bibsonomy.org/bibtex/222e658aa05b4784c411cbf58fb02ed19/lepsky},
booktitle = {Data {Intelligence} and {Cognitive} {Informatics}},
doi = {10.1007/978-981-15-8530-2_42},
editor = {Jeena Jacob, I. and Kolandapalayam Shanmugam, Selvanayaki and Piramuthu, Selwyn and Falkowski-Gilski, Przemyslaw},
file = {Springer Full Text PDF:/Users/le/Zotero/storage/S8BUR9PQ/Satvika et al. - 2021 - A Primer on Word Embedding.pdf:application/pdf},
interhash = {0fd05cba4e7f304ba176eedcd79fe979},
intrahash = {22e658aa05b4784c411cbf58fb02ed19},
isbn = {9789811585302},
keywords = {maschinelles_lernen texttechnologie},
language = {en},
pages = {525--541},
publisher = {Springer},
series = {Algorithms for {Intelligent} {Systems}},
timestamp = {2021-02-08T17:44:00.000+0100},
title = {A primer on word embedding},
year = 2021
}