2022
DOI: 10.1109/access.2022.3225685
|View full text |Cite
|
Sign up to set email alerts
|

A Group Feature Ranking and Selection Method Based on Dimension Reduction Technique in High-Dimensional Data

Abstract: Group feature selection methods select the important group features by removing the irrelevant group features for reducing the complexity of the model. There are a few group feature selection methods that used ranking techniques. Ranking methods provide the relative importance of each group. For this purpose, we developed a sparse group feature ranking method based on the dimension reduction technique for high dimensional data. Firstly, we applied relief to each group to remove irrelevant features. Secondly, w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 32 publications
0
7
0
Order By: Relevance
“…In order to predict GHG emission fluxes based on water quality data, we developed a RF model to describe their relationships. Before constructing the model, feature screening was performed to reduce the number of input features in order to improve the computational efficiency and reliability. The features were sorted according to their importance scores obtained from each prediction, and the one with the lowest score was excluded for the subsequent round of prediction. The model construction process is illustrated in Figure S3.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…In order to predict GHG emission fluxes based on water quality data, we developed a RF model to describe their relationships. Before constructing the model, feature screening was performed to reduce the number of input features in order to improve the computational efficiency and reliability. The features were sorted according to their importance scores obtained from each prediction, and the one with the lowest score was excluded for the subsequent round of prediction. The model construction process is illustrated in Figure S3.…”
Section: Resultsmentioning
confidence: 99%
“…To improve the modeling efficiency, feature selection was performed to exclude the features with relatively low importance scores. In our study, we initially evaluated the prediction performance of the model using 8 input features and recorded the importance score for each feature. The feature importance score reflects the contribution of each feature to the model’s predictions.…”
Section: Methodsmentioning
confidence: 99%
“…Relief is a pre-processing filter that removes irrelevant features by calculating feature weights using proxy statistics. The feature's relevance to the target variable is indicated by its feature score, which runs from −1 to +1 [115]. Nevertheless, it does not handle redundancy and is limited to two-class classification problems.…”
Section: Relief-fmentioning
confidence: 99%
“…Gene expression data, on the other hand, include high dimensionalities that are unimportant while looking for diseases. We cannot trust the high-dimensional gene expression data since it contains redundant information and is thus useless [4]. The inability to process all data at once, as well as the possibility that its subset data processing may lead to the reasons behind over-fitting, information loss, and other such issues, can all be briefly described as limitations of microarray data that are directly affecting the accuracy of classification [5,6].…”
Section: International Journal On Recent and Innovation Trends In Com...mentioning
confidence: 99%