Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014) 2014
DOI: 10.3115/v1/s14-2011
|View full text |Cite
|
Sign up to set email alerts
|

AI-KU: Using Co-Occurrence Modeling for Semantic Similarity

Abstract: In this paper, we describe our unsupervised method submitted to the Cross-Level Semantic Similarity task in Semeval 2014 that computes semantic similarity between two different sized text fragments. Our method models each text fragment by using the cooccurrence statistics of either occurred words or their substitutes. The co-occurrence modeling step provides dense, low-dimensional embedding for each fragment which allows us to calculate semantic similarity using various similarity metrics. Although our current… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 12 publications
0
1
0
Order By: Relevance
“…Although their applications have some differences from the system described in this paper, we consider them relevant because they deal with semantic similarity. (Başkaya, 2014) uses Vector Space Models which have some similarity to our usage of word2vec centroid metrics with the difference that we do not organize the whole text according to the structure of the result matrix, as the VSMs do. The cosine similarity is common for both systems.…”
Section: Related Workmentioning
confidence: 99%
“…Although their applications have some differences from the system described in this paper, we consider them relevant because they deal with semantic similarity. (Başkaya, 2014) uses Vector Space Models which have some similarity to our usage of word2vec centroid metrics with the difference that we do not organize the whole text according to the structure of the result matrix, as the VSMs do. The cosine similarity is common for both systems.…”
Section: Related Workmentioning
confidence: 99%
“…Goldberg (2018, 2019) show how contextualized embeddings can be used for achieving state-of-the-art WSI results. The core idea of their WSI algorithm is based on the intuition, first proposed by Başkaya et al (2013), that occurrences of a word that share a sense, also share in-context substitutes. An MLM is then used to derive top-k word substitutes for each word, and these substitutevectors are clustered to derive word senses.…”
Section: Introductionmentioning
confidence: 99%