Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2016
DOI: 10.18653/v1/n16-1009
|View full text |Cite
|
Sign up to set email alerts
|

Entity-balanced Gaussian pLSA for Automated Comparison

Abstract: Community created content (e.g., product descriptions, reviews) typically discusses one entity at a time and it can be hard as well as time consuming for a user to compare two or more entities. In response, we define a novel task of automatically generating entity comparisons from text. Our output is a table that semantically clusters descriptive phrases about entities. Our clustering algorithm is a Gaussian extension of probabilistic latent semantic analysis (pLSA), in which each phrase is represented in word… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
2
2
1

Relationship

3
2

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 20 publications
0
3
0
Order By: Relevance
“…Probabilistic embeddings have been applied in tasks for building better word representations [3,34], entity comparison [10], facial recognition [7], pose estimation [44], generating multimodal embeddings [2,9], etc. The motivation in some of these tasks is similar to ours -for instance, Qian et al [34] use Gaussian embeddings to represent words to better capture meaning and ambiguity.…”
Section: Probabilistic Embeddingsmentioning
confidence: 99%
“…Probabilistic embeddings have been applied in tasks for building better word representations [3,34], entity comparison [10], facial recognition [7], pose estimation [44], generating multimodal embeddings [2,9], etc. The motivation in some of these tasks is similar to ours -for instance, Qian et al [34] use Gaussian embeddings to represent words to better capture meaning and ambiguity.…”
Section: Probabilistic Embeddingsmentioning
confidence: 99%
“…(b) Hand-designed entity.type and entity.attr specific features. These include indicators for guessing potential types, based on targets of WH ( what , where , which ) words and certain verb classes; multi-sentence features that are based on dependency parses of individual sentences that aid in attribute detection—for example, for every noun and adjective, an attribute indicator feature is on if any of its ancestors is a potential type as indicated by the type feature; indicator features for descriptive phrases (Contractor et al 2016), such as adjective–noun pairs. (c) For each token, we include cluster ids generated from a clustering of word2vec vectors (Mikolov et al 2013) run over a large tourism corpus.…”
Section: Mseq Semantic Labelingmentioning
confidence: 99%
“…(a) Lexical features for capitalization, indicating numerals etc., token-level features based on POS and NER (b) hand-designed x.type and x.attribute specific features. These include indicators for guessing potential types, based on targets of WH (what, where, which) words and certain verb classes; dependency parse features that aid in attribute detection, e.g., for every noun and adjective, an attribute indicator feature is on if any of its ancestors is a potential type as indicated by type feature; indicator features for descriptive phrases (Contractor et al, 2016), such as adjective-noun pairs. (c) For each token, we include cluster ids generated from a clustering of word2vec vectors (Mikolov et al, 2013) run over a large tourism corpus.…”
Section: Supervised Labelingmentioning
confidence: 99%