2013
DOI: 10.1007/978-3-642-40991-2_25
|View full text |Cite
|
Sign up to set email alerts
|

Probabilistic Clustering for Hierarchical Multi-Label Classification of Protein Functions

Abstract: Abstract. Hierarchical Multi-Label Classification is a complex classification problem where the classes are hierarchically structured. This task is very common in protein function prediction, where each protein can have more than one function, which in turn can have more than one sub-function. In this paper, we propose a novel hierarchical multi-label classification algorithm for protein function prediction, namely HMC-PC. It is based on probabilistic clustering, and it makes use of cluster membership probabil… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 14 publications
(6 citation statements)
references
References 23 publications
0
6
0
Order By: Relevance
“…As its name suggests, the algorithm adaptation strategy consists of adapting a traditional algorithm to handle hierarchical constraints. Masera and Blanzieri [6] created a neural network whose architecture incorporates the underlying hierarchy, making gradient updates flow from the neurons associated to the leaves up neurons associated to their parent nodes; Sun et al [8] proposed to use Partial Least Squares to reduce both label and feature dimension, followed by an optimal path selection algorithm; Barros et al [17] proposed a centroid based method where the training data is initially clustered, then predictions are performed by measuring the distance between the new instance and all clusters, the label set associated to the closest cluster is given as the prediction; Borges and Nievola [31] developed a competitive neural network whose architecture replicates the hierarchy; Vens et al [2] also proposed to train a single Predictive Clustering Tree for the entire hierarchy; as an extension of [2], Schietgat et al [21] proposed to use ensemble of Predictive Clustering Trees; Stojanova et al [18] proposed a slight modification for Predictive Clustering Trees in which the correlation between the proteins is also used to build the tree.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…As its name suggests, the algorithm adaptation strategy consists of adapting a traditional algorithm to handle hierarchical constraints. Masera and Blanzieri [6] created a neural network whose architecture incorporates the underlying hierarchy, making gradient updates flow from the neurons associated to the leaves up neurons associated to their parent nodes; Sun et al [8] proposed to use Partial Least Squares to reduce both label and feature dimension, followed by an optimal path selection algorithm; Barros et al [17] proposed a centroid based method where the training data is initially clustered, then predictions are performed by measuring the distance between the new instance and all clusters, the label set associated to the closest cluster is given as the prediction; Borges and Nievola [31] developed a competitive neural network whose architecture replicates the hierarchy; Vens et al [2] also proposed to train a single Predictive Clustering Tree for the entire hierarchy; as an extension of [2], Schietgat et al [21] proposed to use ensemble of Predictive Clustering Trees; Stojanova et al [18] proposed a slight modification for Predictive Clustering Trees in which the correlation between the proteins is also used to build the tree.…”
Section: Related Workmentioning
confidence: 99%
“…Some work, such as [5,10,11,13,17,22], have also decided to not include them. Table 12 presents the datasets evaluated in this work.…”
Section: Evaluated Datasetsmentioning
confidence: 99%
“…Alaydie et al (2012) proposed a boosting-based method for HMC, where at each iteration the label hierarchy is used to select the training set for each classifier. Recently, Barros et al proposed a method for HMC based on the probabilistic clustering with expectation-maximization algorithm (Barros et al 2013).…”
Section: Related Workmentioning
confidence: 99%
“…There has also been recent interest in multi-label text classification [51], [52], [8], [53], [54] Recently, progress has also been made in incorporating hierarchical knowledge to single label classifiers to add additional semantics to the models' learning capabilities such that even when the model makes mistakes dues to data ambiguities, it is able to make semantically better mistakes. Hierarchical information is important in many other applications such as food recognition [55], [25], protein function prediction [6], [7], [56], [57], [58], [59] , image annotation [60], text classification [61], [62], [63]. Some major approaches include imposing logical constraints [4], using hyperbolic embeddings [64], prototype learning [14], label smearing and soft labels, loss modifications [3], multiple learning heads for different levels of the hierarchy [5], hierarchical post-processing [65] and others [66], [67], [68].…”
Section: Related Workmentioning
confidence: 99%