2005
DOI: 10.1243/095440605x31571
|View full text |Cite
|
Sign up to set email alerts
|

Online Discretization of Continuous-Valued Attributes in Rule Induction

Abstract: Machine learning algorithms designed for engineering applications must be able to handle numerical attributes, particularly attributes with real (or continuous) values. Many algorithms deal with continuous-valued attributes by discretizing them before starting the learning process. This paper describes a new approach for discretization of continuousvalued attributes during the learning process. Incorporating discretization within the learning process has the advantage of taking into account the bias inherent i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2005
2005
2016
2016

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 37 publications
0
3
0
Order By: Relevance
“…Search-space pruning rules employed by RULES-6 r′ is any specialisation of rule r and Prune (r) indicates that the children of r should not be searched. The experimental results of many studies [22,23] have indicated that the choice of a discretisation method depends on both the data to be discretised and the learning algorithm. The performance of the four discretisation methods mentioned above when used with the RULES-6 algorithm was evaluated empirically [17].…”
Section: Discretisation Methodsmentioning
confidence: 99%
“…Search-space pruning rules employed by RULES-6 r′ is any specialisation of rule r and Prune (r) indicates that the children of r should not be searched. The experimental results of many studies [22,23] have indicated that the choice of a discretisation method depends on both the data to be discretised and the learning algorithm. The performance of the four discretisation methods mentioned above when used with the RULES-6 algorithm was evaluated empirically [17].…”
Section: Discretisation Methodsmentioning
confidence: 99%
“…After consistent and clean data sets have been formed, data sampling and feature selection techniques are usually employed to reduce the data, thus speeding up the data mining process. Data often contain a mixture of categorical and continuous-valued attributes, and therefore continuous-valued attributes may have to be discretized first [35]. Data preparation is the most time-consuming stage in the whole data mining process.…”
Section: Overview Of Data Miningmentioning
confidence: 99%
“…39 Instead of examining all individual values, this method examines only the boundary values of each numeric feature during learning. Split points are added when adjacent values of the same feature are identified, where each belongs to a different class.…”
mentioning
confidence: 99%