Proceedings of the 2018 World Wide Web Conference on World Wide Web - WWW '18 2018
DOI: 10.1145/3178876.3185998
|View full text |Cite
|
Sign up to set email alerts
|

Parabel

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
28
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 138 publications
(28 citation statements)
references
References 18 publications
0
28
0
Order By: Relevance
“…In contrast, the algorithmic complexity of kNN scales logarithmically with the number of training points used. For this reason, tree-based methods are common in other Extreme Classification applications [9], where similarly massive numbers of categories are under consideration. Unfortunately, the resulting faster runtimes come at a significant cost; kNN often gives far worse results in practice than a more rigorous Bayesian approach.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…In contrast, the algorithmic complexity of kNN scales logarithmically with the number of training points used. For this reason, tree-based methods are common in other Extreme Classification applications [9], where similarly massive numbers of categories are under consideration. Unfortunately, the resulting faster runtimes come at a significant cost; kNN often gives far worse results in practice than a more rigorous Bayesian approach.…”
Section: Methodsmentioning
confidence: 99%
“…For example, in the human proteome, there are about 20,000 proteins, which when processed with an amino-acid specific protease such as trypsin can correspond to hundreds of thousands or even millions of distinct peptides, each of which can potentially vary due to post-translational modifications or experiment-specific processing. This puts fluorosequencing data analysis squarely in the realm of Extreme Classification problems, which are known to be challenging to handle in practice [9].…”
Section: Introductionmentioning
confidence: 99%
“…This is a basic approach, focusing on improving model efficiency, but requiring an extremely large amount of computational complexity, typically represented by DiSMEC [5] and PPDSparse [6]; 2) The main intention of Tree-based methods is to overcome the high computational complexity of the One-vs-all method. It is based on the idea of performing hierarchical classification, where the features of the labels can be implemented in clusters, represented by Parabel [7]; 3) Embedding-based method approaches that perform high-and low-dimensional spatial transformations through label compression and decompression, sacrificing a certain amount of information in exchange for a reduction in computational complexity, represented by AnnexML [8]; 4) Deep learning method is the current mainstream direction. Since 2017, XML-CNN has become the first successful deep learning method in the XMC domain and introduced the important concept of bottleneck layer.…”
Section: Related Workmentioning
confidence: 99%
“…We compare the most representative methods as our baselines including: DiSMEC [5] is a largescale distributed framework for learning 1-vs-All approaches. Parabel [7], a method using PLTs, can be considered as taking the advantage of both tree-based and 1-vs-All methods. Bonsai [16], another method using PLTs, is similar to Parabel.…”
Section: Baselinesmentioning
confidence: 99%
“…Many machine learning based systems for automated subject indexing have been developed since the 1990s, when the approach became the dominant paradigm for automated subject indexing (Sebastiani, 2002). Some recent examples for which an implementation is available include Magpie (Berger, 2015;Kim, 2014), FastXML (Prabhu & Varma, 2014), PD-Sparse (Yen, Huang, Zhong, Ravikumar, & Dhillon, 2016), fastText (Joulin, Grave, Bojanowski, & Mikolov, 2017), Quadflor (Galke, Mai, Schelten, Brunsch, & Scherp, 2017), AnnexML (Tagami, 2017) and Parabel (Prabhu, Kag, Harsola, Agrawal, & Varma, 2018).…”
Section: Machine Learning Approachesmentioning
confidence: 99%