2020
DOI: 10.1609/aaai.v34i05.6292
|View full text |Cite
|
Sign up to set email alerts
|

P-SIF: Document Embeddings Using Partition Averaging

Abstract: Simple weighted averaging of word vectors often yields effective representations for sentences which outperform sophisticated seq2seq neural models in many tasks. While it is desirable to use the same method to represent documents as well, unfortunately, the effectiveness is lost when representing long documents involving multiple sentences. One of the key reasons is that a longer document is likely to contain words from many different topics; hence, creating a single vector while ignoring all the topical stru… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
5

Relationship

1
9

Authors

Journals

citations
Cited by 18 publications
(13 citation statements)
references
References 18 publications
(38 reference statements)
0
13
0
Order By: Relevance
“…In our work, we adhere to the linear evaluation protocol and use a simple logistic regression classifier in evaluating word representations (Arora et al, 2019;Gupta et al, 2020), whereas (Mu and Viswanath, 2018) use a neural network as their classifier. The red colored cells in Table 5 denote the cases where the original embeddings outperformed their Post Processed (PPA) counterparts.…”
Section: Sentence Classification Tasksmentioning
confidence: 99%
“…In our work, we adhere to the linear evaluation protocol and use a simple logistic regression classifier in evaluating word representations (Arora et al, 2019;Gupta et al, 2020), whereas (Mu and Viswanath, 2018) use a neural network as their classifier. The red colored cells in Table 5 denote the cases where the original embeddings outperformed their Post Processed (PPA) counterparts.…”
Section: Sentence Classification Tasksmentioning
confidence: 99%
“…Although we focused on re-evaluating WMD, our findings are beneficial for other topics as well. First, the datasets released by the WMD paper have been used in non-OT studies as well [72,50,31]. Our findings regarding the duplicate documents are important for proper evaluations of other methods as well.…”
Section: Discussionmentioning
confidence: 99%
“…First, we use word weights in constructing semantically similar groups and can incorporate different weighting scheme in our model, such as SIF [11], GEM [12]. Second, different clustering schemes such as the Gaussian mixture model and dictionary learning can be utilized to construct semantically similar groups [13], [31]. Finally, the intra-group descriptor can be replaced by methods like VLAWE [18] and p-means [14].…”
Section: Discussionmentioning
confidence: 99%