2008
DOI: 10.1007/s10994-008-5089-z
|View full text |Cite
|
Sign up to set email alerts
|

gBoost: a mathematical programming approach to graph classification and regression

Abstract: Graph mining methods enumerate frequently appearing subgraph patterns, which can be used as features for subsequent classification or regression. However, frequent patterns are not necessarily informative for the given learning problem. We propose a mathematical programming boosting method (gBoost) that progressively collects informative patterns. Compared to AdaBoost, gBoost can build the prediction rule with fewer iterations. To apply the boosting method to graph data, a branch-and-bound pattern search algor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
187
0

Year Published

2009
2009
2024
2024

Publication Types

Select...
6
3

Relationship

2
7

Authors

Journals

citations
Cited by 125 publications
(188 citation statements)
references
References 34 publications
1
187
0
Order By: Relevance
“…who report a best AUROC of 88.9% using gBoost, 43 and an accuracy of 82.5%. Kazius et al 33 achieved a training accuracy of 83% with a manually constructed model, using all data without cross-validation.…”
Section: Resultsmentioning
confidence: 98%
“…who report a best AUROC of 88.9% using gBoost, 43 and an accuracy of 82.5%. Kazius et al 33 achieved a training accuracy of 83% with a manually constructed model, using all data without cross-validation.…”
Section: Resultsmentioning
confidence: 98%
“…Computing the edit distance between two objects can be done in polynomial time using Dynamic Programming, [232]. More complex structures, such as trees, or, even more generally, graphs, appear naturally in many real world data sets of biological sequences, or semi-structured texts such as HTML or XML, [206,126,246]. Computing in these cases a dissimilarity δ for a given pair of objects may lead to hard combinatorial optimization problems, [27].…”
Section: Nearest-neighbors Methodsmentioning
confidence: 99%
“…Optimization has shown to be useful to decide how classifiers should be ensembled. For instance, in [77,206] a column generation approach, [105], is used in the boosting environment, whereas a quadratic programming model is used in [174].…”
Section: Classification Treesmentioning
confidence: 99%
“…More recently, substructure boosting approach has been successfully applied to different learning tasks on various kinds of data including RNA secondary structure clustering [9], video classification [10], and QSAR [11], [12]. These methods combine statistical learning algorithms with pattern mining algorithms to directly mine discriminative patterns which are optimal for the subsequent learning task in an iterative fashion [13].…”
Section: Introductionmentioning
confidence: 99%