Abstract:Frequency is one of the major factors for training quality word embeddings. Several studies have recently discussed the stability of word embeddings in general domain and suggested factors influencing the stability. In this work, we conduct a detailed analysis on the stability of concept embeddings in medical domain, particularly in relations with concept frequency. The analysis reveals the surprising high stability of low‐frequency concepts: low‐frequency (<100) concepts have the same high stability as high‐f… Show more
“…• Wikipedia (1.5B) [22] • NYT(58M) and Europarl (61M) [32] • Brown, Project Gutenberg and Reuters (10k each) [8] • US Federal Courts of Appeals (38k), NYT (22k) and Reddit (26k) [1] • NIPS between 2007 and 2012 (2M) [11] • Ohsumed dataset (34M) [21] • Google N-gram cor-pus: English Fiction(4.8B) and German(0.7B) [18] • BNC and ACL An-thology Reference corpus [28] Wikipedia, News-Crawl (2007), Lyrics and Europarl (50M each)…”
Section: Previous Work Our Workmentioning
confidence: 99%
“…• Medium frequency words have the lowest reliability [18] • Frequency is not a major factor in stability [32] • Words having a low and high frequency range have a tendency to display more variation. Medium frequency words show more stability [28] • Frequency does not directly affect the stability of medical word embeddings [21] •All word groups(low, medium, and high frequency) show high variance in word stability.…”
Section: Word Frequencymentioning
confidence: 99%
“…Stability evaluation of WEMs has received a lot of attention in the recent past [1,8,11,21,22,32]. Such evaluation is crucial because word embeddings are widely used today for a variety of NLP tasks such as sentiment analysis [23,30], named entity recognition [20,29] and part-of-speech (POS) tagging [31,32].…”
A representation learning method is considered stable if it consistently generates similar representation of the given data across multiple runs. Word Embedding Methods (WEMs) are a class of representation learning methods that generate dense vector representation for each word in the given text data. The central idea of this paper is to explore the stability measurement of WEMs using intrinsic evaluation based on word similarity. We experiment with three popular WEMs: Word2Vec, GloVe, and fastText. For stability measurement, we investigate the effect of five parameters involved in training these models. We perform experiments using four real-world datasets from different domains: Wikipedia, News, Song lyrics, and European parliament proceedings. We also observe the effect of WEM stability on three downstream tasks: Clustering, POS tagging, and Fairness evaluation. Our experiments indicate that amongst the three WEMs, fastText is the most stable, followed by GloVe and Word2Vec.
“…• Wikipedia (1.5B) [22] • NYT(58M) and Europarl (61M) [32] • Brown, Project Gutenberg and Reuters (10k each) [8] • US Federal Courts of Appeals (38k), NYT (22k) and Reddit (26k) [1] • NIPS between 2007 and 2012 (2M) [11] • Ohsumed dataset (34M) [21] • Google N-gram cor-pus: English Fiction(4.8B) and German(0.7B) [18] • BNC and ACL An-thology Reference corpus [28] Wikipedia, News-Crawl (2007), Lyrics and Europarl (50M each)…”
Section: Previous Work Our Workmentioning
confidence: 99%
“…• Medium frequency words have the lowest reliability [18] • Frequency is not a major factor in stability [32] • Words having a low and high frequency range have a tendency to display more variation. Medium frequency words show more stability [28] • Frequency does not directly affect the stability of medical word embeddings [21] •All word groups(low, medium, and high frequency) show high variance in word stability.…”
Section: Word Frequencymentioning
confidence: 99%
“…Stability evaluation of WEMs has received a lot of attention in the recent past [1,8,11,21,22,32]. Such evaluation is crucial because word embeddings are widely used today for a variety of NLP tasks such as sentiment analysis [23,30], named entity recognition [20,29] and part-of-speech (POS) tagging [31,32].…”
A representation learning method is considered stable if it consistently generates similar representation of the given data across multiple runs. Word Embedding Methods (WEMs) are a class of representation learning methods that generate dense vector representation for each word in the given text data. The central idea of this paper is to explore the stability measurement of WEMs using intrinsic evaluation based on word similarity. We experiment with three popular WEMs: Word2Vec, GloVe, and fastText. For stability measurement, we investigate the effect of five parameters involved in training these models. We perform experiments using four real-world datasets from different domains: Wikipedia, News, Song lyrics, and European parliament proceedings. We also observe the effect of WEM stability on three downstream tasks: Clustering, POS tagging, and Fairness evaluation. Our experiments indicate that amongst the three WEMs, fastText is the most stable, followed by GloVe and Word2Vec.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.