2016
DOI: 10.1007/s10115-016-0937-9
|View full text |Cite
|
Sign up to set email alerts
|

Selective AnDE for large data learning: a low-bias memory constrained approach

Abstract: Learning from data that are too big to fit into memory poses great challenges to currently available learning approaches. Averaged n-Dependence Estimators (AnDE) allows for a flexible learning from out-of-core data, by varying the value of n (number of super parents). Hence, AnDE is especially appropriate for learning from large quantities of data. Memory requirement in AnDE, however, increases combinatorially with the number of attributes and the parameter n. In large data learning, number of attributes is of… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
7
2

Relationship

3
6

Authors

Journals

citations
Cited by 16 publications
(6 citation statements)
references
References 17 publications
0
6
0
Order By: Relevance
“…Then, in the prediction phase, the A2DE classifier predicts the class by even outsourcing one dependence estimator to the number of records available in the dataset. Further, the low bias characteristics and single-pass learning through training data of A2DE make it popular for big data analytics with high accuracy [47], linear time complexity but with high variance, and increased space complexity [48].…”
Section: Ande (Averaged N-dependence Estimator)mentioning
confidence: 99%
“…Then, in the prediction phase, the A2DE classifier predicts the class by even outsourcing one dependence estimator to the number of records available in the dataset. Further, the low bias characteristics and single-pass learning through training data of A2DE make it popular for big data analytics with high accuracy [47], linear time complexity but with high variance, and increased space complexity [48].…”
Section: Ande (Averaged N-dependence Estimator)mentioning
confidence: 99%
“…The functional domain of one single classifier may be limited as a result of ignoring the dependencies between some attributes. Classifiers that use the forest or ensemble method are commonly applied to fill the gap [12,14,15]. In the following subsection, we first introduce NB and its corresponding ensemble classifier, that is, AODE.…”
Section: Bayesian Network Classifiersmentioning
confidence: 99%
“…After the discovery of NB, many state-of-the-art algorithms, for example, tree-augmented naive Bayes (TAN) [10] and a k-dependence Bayesian classifier (KDB) [11], are proposed to relax the independence assumption by allowing conditional dependence between attributes X i and X j , which is measured by conditional mutual information I(X i ; X j |C). In order to improve predictive accuracy relative to a single model, ensemble methods [12,13], for example, averaged one-dependence estimator (AODE) [14] and averaged tree-augmented naive Bayes (ATAN) [15] methods, generate multiple global models from a single learning algorithm through randomization (or perturbation). The KDB is a form of a restricted Bayesian network classifier with numerous desirable properties in the context of learning from large quantities of data.…”
Section: Introductionmentioning
confidence: 99%
“…Although information theory was primarily concerned with the problem of digital communication when it was first introduced by Claude E. Shannon in the 1940s [15], the theory has much broader applicability in the field of classification [16,17]. Here, we review several commonly used definitions.…”
Section: Information Theorymentioning
confidence: 99%