Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014) 2014
DOI: 10.3115/v1/s14-2039
|View full text |Cite
|
Sign up to set email alerts
|

DLS$@$CU: Sentence Similarity from Word Alignment

Abstract: We present an algorithm for computing the semantic similarity between two sentences. It adopts the hypothesis that semantic similarity is a monotonically increasing function of the degree to which (1) the two sentences contain similar semantic units, and (2) such units occur in similar semantic contexts. With a simplistic operationalization of the notion of semantic units with individual words, we experimentally show that this hypothesis can lead to state-of-the-art results for sentencelevel semantic similarit… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 57 publications
(21 citation statements)
references
References 4 publications
0
21
0
Order By: Relevance
“…This is because the similarities between short and grammatically simple sentences of MSRvid-12 highly depend on the syntactic roles the concepts occupy (e.g., "The man is playing with the dog" vs. "The dog is playing with the man"). Whereas the model of Sultan et al (2014), using dependency parsing between sentences can take syntactic roles of concepts into account to better align words between sentences, our simple STS metric -designed to be language-independent and resource-light -cannot.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…This is because the similarities between short and grammatically simple sentences of MSRvid-12 highly depend on the syntactic roles the concepts occupy (e.g., "The man is playing with the dog" vs. "The dog is playing with the man"). Whereas the model of Sultan et al (2014), using dependency parsing between sentences can take syntactic roles of concepts into account to better align words between sentences, our simple STS metric -designed to be language-independent and resource-light -cannot.…”
Section: Resultsmentioning
confidence: 99%
“…The cross-lingual performances of the OptAlign model on the English-Spanish variants of the News-16 and MulSrc-16 datasets come reasonably close to the best performing models (Brychcín and Svoboda, 2016;Jimenez, 2016) from the SemEval 2016 STS cross-lingual evaluation. The 4-5% difference in performance (86% compared to 90% on News-16 and 77% compared to 82% on MulSrc-16) seems rather satisfying considering that OptAlign is very resource-light and language-independent, whereas the best-performing system (Brychcín and Svoboda, 2016) employs a full-blown MT system (Google translate), and a word alignment model that requires a dependency parser, a named entity recognizer, and a paraphrase database (Sultan et al, 2014).…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…The paraphrases in it are extracted automatically from multilingual resources. It is reported useful in many other tasks such as recognizing textual entailment [31], [32], measuring the semantic similarity [33]- [35], monolingual alignment [36], [37], and natural language generation [38].…”
Section: B the Lexicon Layermentioning
confidence: 99%
“…The overall sentence measure depends on the number of tokens, RDF triples that entail the semantic layer. In the same area, [52] combined the words meanings and phrase context in a sentence measure. The meaning words are implied by extracting words' lemma from a dictionary, whereas phrase context usage was extracted using a huge para-phrase alignment database [53].…”
Section: Hybrid-based Methodsmentioning
confidence: 99%