2020
DOI: 10.3126/jiee.v3i1.34327
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Estimation of Nepali Word Representations in Vector Space

Abstract: Word representation is a means of representing a word as mathematical entities that can be read, reasoned and manipulated by computational models. The representation is required for input to any new modern data models and in many cases, the accuracy of a model depends on it. In this paper, we analyze various methods of calculating vector space for Nepali words and postulate a word to vector model based on the Skip-gram model with NCE loss capturing syntactic and semantic word relationships. This is an at… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 62 publications
(12 citation statements)
references
References 1 publication
0
12
0
Order By: Relevance
“…We trained the neural network on a corpus of 1.1 million words sourced from 22 individual blogs and online forums ( Multimedia Appendix 1 ). We used the skip-gram negative sampling variant of the word2vec neural network algorithm described by Mikolov et al [ 9 ] to discover community words and phrases for disease symptoms. Briefly, the neural network model was trained to predict context words that appear in close proximity with symptom keywords in the corpus text.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We trained the neural network on a corpus of 1.1 million words sourced from 22 individual blogs and online forums ( Multimedia Appendix 1 ). We used the skip-gram negative sampling variant of the word2vec neural network algorithm described by Mikolov et al [ 9 ] to discover community words and phrases for disease symptoms. Briefly, the neural network model was trained to predict context words that appear in close proximity with symptom keywords in the corpus text.…”
Section: Methodsmentioning
confidence: 99%
“…We address this limitation with a novel approach based on a neural network, specifically a word embedding [ 9 ], to identify words and phrases that patients with chronic obstructive pulmonary disease (COPD) use to describe their experiences of living with the disease. Unlike traditional neural network approaches, a word embedding is not trained on any specific set of scientific keywords [ 10 , 11 ].…”
Section: Introductionmentioning
confidence: 99%
“…Embedding technique has been developed and immediately obtained a considerable attention and success in ML community in general, and in natural language processing [58] and recommendation systems [59] in particular. This technique allows condensing the dimensions of the input features, thereby it contributes to stabilising the learning process.…”
Section: Stream Representationmentioning
confidence: 99%
“…With the emergence of word vector technology that converts words into numerical vectors, word meaning measurement becomes possible. The main derived word vector generation models include Word2vec [15], G1oVe [16], ELMo [17], and BERT [18]. The most commonly used are the Word2vec model and the BERT model.…”
Section: The Research Statusmentioning
confidence: 99%