Search citation statements
Paper Sections
Citation Types
Year Published
Publication Types
Relationship
Authors
Journals
In this article, we develop and analyze four algorithmspatterns from large databases. As described in Fayyad for feature selection in the context of rough set method- (1996) and Simoudis (1996), this process is typically ology. The initial state and the feasibility criterion of all made up of selection and sampling, preprocessing and these algorithms are the same. That is, they start with a cleaning, transformation and reduction, data mining, and given feature set and progressively remove features, evaluation steps. The first step in the data-mining process while controlling the amount of degradation in classification quality. These algorithms, however, differ in the is to select a target data set from a database (or a data heuristics used for pruning the search space of features. warehouse) and to possibly sample the target data. The Our experimental results confirm the expected relationpreprocessing and data cleaning step handles noise and ship between the time complexity of these algorithms unknown values, as well as accounting for missing data and the classification accuracy of the resulting upper fields, time sequence information, and so forth. The data classifiers. Our experiments demonstrate that a u-reduct of a given feature set can be found efficiently. Although reduction and transformation step involves finding relewe have adopted upper classifiers in our investigations, vant features depending on the goal of the task and certain the algorithms presented can, however, be used with transformations on the data such as converting one type any method of deriving a classifier, where the quality of of data to another (e.g., changing nominal values into classification is a monotonically decreasing function of the size of the feature set. We compare the performance numeric ones, discretizing continuous values), and/or deof upper classifiers with those of lower classifiers. We fining new attributes. In the mining step, the user may find that upper classifiers perform better than lower apply one or more knowledge discovery techniques on classifiers for a duodenal ulcer data set. This should be the transformed data to extract valuable patterns. Finally, generally true when there is a small number of elements the evaluation step involves interpreting the result (or in the boundary region. An upper classifier has some important features that make it suitable for data mining discovered pattern) with respect to the goal/task at hand. applications. In particular, we have shown that the upperNote that the data-mining process is not linear and inclassifiers can be summarized at a desired level of abvolves a variety of feedback loops, because any one step straction by using extended decision tables. We also can result in changes in preceding or succeeding steps.point out that an upper classifier results in an inconsistent decision algorithm, which can be interpreted deter-Furthermore, the nature of a large, real-world data set, ministically or non-deterministically to obtain a consiswhich may contain noisy, incomplete, d...
In this article, we develop and analyze four algorithmspatterns from large databases. As described in Fayyad for feature selection in the context of rough set method- (1996) and Simoudis (1996), this process is typically ology. The initial state and the feasibility criterion of all made up of selection and sampling, preprocessing and these algorithms are the same. That is, they start with a cleaning, transformation and reduction, data mining, and given feature set and progressively remove features, evaluation steps. The first step in the data-mining process while controlling the amount of degradation in classification quality. These algorithms, however, differ in the is to select a target data set from a database (or a data heuristics used for pruning the search space of features. warehouse) and to possibly sample the target data. The Our experimental results confirm the expected relationpreprocessing and data cleaning step handles noise and ship between the time complexity of these algorithms unknown values, as well as accounting for missing data and the classification accuracy of the resulting upper fields, time sequence information, and so forth. The data classifiers. Our experiments demonstrate that a u-reduct of a given feature set can be found efficiently. Although reduction and transformation step involves finding relewe have adopted upper classifiers in our investigations, vant features depending on the goal of the task and certain the algorithms presented can, however, be used with transformations on the data such as converting one type any method of deriving a classifier, where the quality of of data to another (e.g., changing nominal values into classification is a monotonically decreasing function of the size of the feature set. We compare the performance numeric ones, discretizing continuous values), and/or deof upper classifiers with those of lower classifiers. We fining new attributes. In the mining step, the user may find that upper classifiers perform better than lower apply one or more knowledge discovery techniques on classifiers for a duodenal ulcer data set. This should be the transformed data to extract valuable patterns. Finally, generally true when there is a small number of elements the evaluation step involves interpreting the result (or in the boundary region. An upper classifier has some important features that make it suitable for data mining discovered pattern) with respect to the goal/task at hand. applications. In particular, we have shown that the upperNote that the data-mining process is not linear and inclassifiers can be summarized at a desired level of abvolves a variety of feedback loops, because any one step straction by using extended decision tables. We also can result in changes in preceding or succeeding steps.point out that an upper classifier results in an inconsistent decision algorithm, which can be interpreted deter-Furthermore, the nature of a large, real-world data set, ministically or non-deterministically to obtain a consiswhich may contain noisy, incomplete, d...
In this article, we develop and analyze four algorithmspatterns from large databases. As described in Fayyad for feature selection in the context of rough set method- (1996) and Simoudis (1996), this process is typically ology. The initial state and the feasibility criterion of all made up of selection and sampling, preprocessing and these algorithms are the same. That is, they start with a cleaning, transformation and reduction, data mining, and given feature set and progressively remove features, evaluation steps. The first step in the data-mining process while controlling the amount of degradation in classification quality. These algorithms, however, differ in the is to select a target data set from a database (or a data heuristics used for pruning the search space of features. warehouse) and to possibly sample the target data. The Our experimental results confirm the expected relationpreprocessing and data cleaning step handles noise and ship between the time complexity of these algorithms unknown values, as well as accounting for missing data and the classification accuracy of the resulting upper fields, time sequence information, and so forth. The data classifiers. Our experiments demonstrate that a u-reduct of a given feature set can be found efficiently. Although reduction and transformation step involves finding relewe have adopted upper classifiers in our investigations, vant features depending on the goal of the task and certain the algorithms presented can, however, be used with transformations on the data such as converting one type any method of deriving a classifier, where the quality of of data to another (e.g., changing nominal values into classification is a monotonically decreasing function of the size of the feature set. We compare the performance numeric ones, discretizing continuous values), and/or deof upper classifiers with those of lower classifiers. We fining new attributes. In the mining step, the user may find that upper classifiers perform better than lower apply one or more knowledge discovery techniques on classifiers for a duodenal ulcer data set. This should be the transformed data to extract valuable patterns. Finally, generally true when there is a small number of elements the evaluation step involves interpreting the result (or in the boundary region. An upper classifier has some important features that make it suitable for data mining discovered pattern) with respect to the goal/task at hand. applications. In particular, we have shown that the upperNote that the data-mining process is not linear and inclassifiers can be summarized at a desired level of abvolves a variety of feedback loops, because any one step straction by using extended decision tables. We also can result in changes in preceding or succeeding steps.point out that an upper classifier results in an inconsistent decision algorithm, which can be interpreted deter-Furthermore, the nature of a large, real-world data set, ministically or non-deterministically to obtain a consiswhich may contain noisy, incomplete, d...
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations鈥揷itations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright 漏 2024 scite LLC. All rights reserved.
Made with 馃挋 for researchers
Part of the Research Solutions Family.