2020
DOI: 10.1016/j.knosys.2020.106346
|View full text |Cite
|
Sign up to set email alerts
|

Novel metrics for computing semantic similarity with sense embeddings

Abstract: In the last years many efforts have been spent to build word embeddings, a representational device in which word meanings are described through dense unit vectors of real numbers over a continuous, high-dimensional Euclidean space, where similarity can be interpreted as a metric. Afterwards, senselevel embeddings have been proposed to describe the meaning of senses, rather than terms. More recently, additional intermediate representations have been designed, providing a vector description for pairs ⟨term, sens… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(4 citation statements)
references
References 72 publications
(101 reference statements)
0
2
0
Order By: Relevance
“…As with GWCS, LMMS-SP XLNet-L stands out with clearly best results. While LMMS-SP also outperforms most non-contextual embeddings, it still underperforms LessLex (Colla et al, 2020b) embeddings, which are based on ensembles and learned using BabelNet. We also note that LMMS-SP performs particularly well on the 'Observed' set corresponding to senses learned from annotated corpora.…”
Section: Solutionmentioning
confidence: 99%
“…As with GWCS, LMMS-SP XLNet-L stands out with clearly best results. While LMMS-SP also outperforms most non-contextual embeddings, it still underperforms LessLex (Colla et al, 2020b) embeddings, which are based on ensembles and learned using BabelNet. We also note that LMMS-SP performs particularly well on the 'Observed' set corresponding to senses learned from annotated corpora.…”
Section: Solutionmentioning
confidence: 99%
“…Because precision and accuracy will affect the appropriate size and produce the right size results [56]. In addition, a measuring instrument must have high accuracy and precision to reduce the occurrence of errors in measurement [57].…”
Section: Accuracymentioning
confidence: 99%
“…Contextual annotation thus involves a hidden task, that is, the identification of the meanings at the base of the concreteness rating (more on the meaning-identification task can be found in [49]). All targets were then meaning-annotated according to WordNet and are part of the released data.…”
Section: Plos Onementioning
confidence: 99%
“…One main assumption underlying the whole work is that each term may have more associated meanings, and meaning selection is determined based on the context, which, in turn, corresponds to some sort of word meaning disambiguation. That is, the concreteness rating involves a hidden task in which meanings are identified [49], and concreteness is supposed to be a property of word meanings rather than of word forms/terms. While we fully acknowledge the importance of addressing word ambiguity, encompassing both homonymy (where a word form's meanings have distinct historical origins and are not related) and polysemy (where a word form's meanings stem from the same lexical source and are related), practical considerations challenge the rigid application of the etymological criterion.…”
Section: Introductionmentioning
confidence: 99%