2018
DOI: 10.1007/s10994-018-5718-0
|View full text |Cite
|
Sign up to set email alerts
|

Accurate parameter estimation for Bayesian network classifiers using hierarchical Dirichlet processes

Abstract: This paper introduces a novel parameter estimation method for the probability tables of Bayesian network classifiers (BNCs), using hierarchical Dirichlet processes (HDPs). The main result of this paper is to show that improved parameter estimation allows BNCs to outperform leading learning methods such as Random Forest for both 0-1 loss and RMSE, albeit just on categorical datasets.As data assets become larger, entering the hyped world of "big", efficient accurate classification requires three main elements: (… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0
2

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 19 publications
(13 citation statements)
references
References 28 publications
(40 reference statements)
0
11
0
2
Order By: Relevance
“…Although more sophisticated smoothing methods such as Kneser-Ney [8] and Modified Kneser-Ney [2] have been used in language modelling for a long time, M-branch was the first hierarchical smoothing method for decision trees proposed in 2003 [6]. A recent smoothing method called Hierarchical Dirichlet Process (HDP) has had great success on language modelling [17] and Bayesian Network Classifiers [13], whereas it has not been used on decision trees. The following part introduces these methods in detail.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Although more sophisticated smoothing methods such as Kneser-Ney [8] and Modified Kneser-Ney [2] have been used in language modelling for a long time, M-branch was the first hierarchical smoothing method for decision trees proposed in 2003 [6]. A recent smoothing method called Hierarchical Dirichlet Process (HDP) has had great success on language modelling [17] and Bayesian Network Classifiers [13], whereas it has not been used on decision trees. The following part introduces these methods in detail.…”
Section: Related Workmentioning
confidence: 99%
“…Equation 3 can also be explained by the Hierarchical Chinese Restaurant Process (CRP) [19]. Please refer to [13] for more detail of HDP smoothing on Bayesian Network Classifiers.…”
Section: Hdp Smoothingmentioning
confidence: 99%
See 2 more Smart Citations
“…n ] is the observation over n random variables: x 1 ∼ X 1 , · · · , x n ∼ X n . Under this assumption, a Bayesian network can be formally described by B =< G, Θ G >, where G is a directed acyclic graph and Θ G the set of parameters that can maximize the likelihood [7,23]. The i-th node in G corresponds to a random variable X i , and an edge between two connected nodes indicates the direct dependency.…”
Section: Inference Graphmentioning
confidence: 99%