2006
DOI: 10.1007/11941439_28
|View full text |Cite
|
Sign up to set email alerts
|

Voting Massive Collections of Bayesian Network Classifiers for Data Streams

Abstract: Abstract. We present a new method for voting exponential (in the number of attributes) size sets of Bayesian classifiers in polynomial time with polynomial memory requirements. Training is linear in the number of instances in the dataset and can be performed incrementally. This allows the collection to learn from massive data streams. The method allows for flexibility in balancing computational complexity, memory requirements and classification performance. Unlike many other incremental Bayesian methods, all s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
31
0
2

Year Published

2013
2013
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 27 publications
(33 citation statements)
references
References 4 publications
0
31
0
2
Order By: Relevance
“…Generally, all of the sequential and structural features are integrated to construct comprehensive feature representations of query proteins. For prediction engine construction, they build an ensemble classifier model, which fuses five basic classifier models (RF [30], NB [31], Bayes Net [32], LibSVM [33], and SMO (Sequential Minimal Optimization) [34]) with an average probability strategy. Importantly, an online webserver that implements the PFPA method is developed and freely available at .…”
Section: Recent Representative Methods For Protein Fold Recognitionmentioning
confidence: 99%
“…Generally, all of the sequential and structural features are integrated to construct comprehensive feature representations of query proteins. For prediction engine construction, they build an ensemble classifier model, which fuses five basic classifier models (RF [30], NB [31], Bayes Net [32], LibSVM [33], and SMO (Sequential Minimal Optimization) [34]) with an average probability strategy. Importantly, an online webserver that implements the PFPA method is developed and freely available at .…”
Section: Recent Representative Methods For Protein Fold Recognitionmentioning
confidence: 99%
“…The conditional probability table for the values of the variables indicate each possible combination of the values of its parent nodes [22]. Training process of Bayesian networks consists of two stages, namely learning a network structure and learning the probability tables [23]. There are several different ways of structure learning, such as local score metrics, conditional independence tests, global score metrics and fixed structure.…”
Section: Classification Algorithmsmentioning
confidence: 99%
“…There are several different ways of structure learning, such as local score metrics, conditional independence tests, global score metrics and fixed structure. Based on these ways, a number of search algorithms, such as hill climbing, simulated annealing and tabu search are implemented in Weka [23].…”
Section: Classification Algorithmsmentioning
confidence: 99%
“…Moreover, the SPODE model (e.g., AODE) has already been improved for the highly scalable attribute problem in [10]. Also, SPODE models can be trained incrementally [5]. When facing huge word histograms, a hashing correlated feature approach in [4] has been proposed to rank the features.…”
Section: Text Categorization Taskmentioning
confidence: 99%