Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2018
DOI: 10.18653/v1/p18-1222
|View full text |Cite
|
Sign up to set email alerts
|

hyperdoc2vec: Distributed Representations of Hypertext Documents

Abstract: Hypertext documents, such as web pages and academic papers, are of great importance in delivering information in our daily life. Although being effective on plain documents, conventional text embedding methods suffer from information loss if directly adapted to hyper-documents. In this paper, we propose a general embedding approach for hyper-documents, namely, hyperdoc2vec, along with four criteria characterizing necessary information that hyper-document embedding models should preserve. Systematic comparisons… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
107
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 27 publications
(107 citation statements)
references
References 25 publications
0
107
0
Order By: Relevance
“…Moreover, high thresholds for the translation probability may be set to make the machine translation approach feasible [74]. (d) Neural networks [48,67,75,91,153,174] This group contains not only many approaches to local citation recommendation (6 out of 17, that is, 35%), but also the most recent ones: here, papers have been published since 2014. Due to the large field of neural network research in general, the architectures proposed here also vary considerably.…”
Section: Comparison Of Local Citation Recommendation Approachesmentioning
confidence: 99%
“…Moreover, high thresholds for the translation probability may be set to make the machine translation approach feasible [74]. (d) Neural networks [48,67,75,91,153,174] This group contains not only many approaches to local citation recommendation (6 out of 17, that is, 35%), but also the most recent ones: here, papers have been published since 2014. Due to the large field of neural network research in general, the architectures proposed here also vary considerably.…”
Section: Comparison Of Local Citation Recommendation Approachesmentioning
confidence: 99%
“…Given a citation relation, DocCit2Vec picks one publication from the citation list as the target, and utilizes the surrounding context and structural context as known knowledge to maximize the occurrence of the target citation by updating the parameters (i.e., the embedding vectors) of the neural network. The model learns two document embedding vectors IN and OUT, where the IN vector d I characterizes the document as a citing paper and the OUT vector d O encodes its role as a cited paper [9]. In addition, the model learns IN word vectors w I .…”
Section: A Representing Documents and Citationsmentioning
confidence: 99%
“…The architecture of DocCit2Vec-avg is constructed based on the pv-dm structure of Doc2Vec [14] and HyperDoc2Vec [9], shown on the right side of {w|w ∈ C }. The output layer is computed using a multiclass softmax classifier, and the output value is regarded as the probability of the occurrence of d t .…”
Section: B Doccit2vec With An Average Hidden Layermentioning
confidence: 99%
See 2 more Smart Citations