2020
DOI: 10.1186/s40537-020-00328-3
|View full text |Cite
|
Sign up to set email alerts
|

Context pre-modeling: an empirical analysis for classification based user-centric context-aware predictive modeling

Abstract: Nowadays, machine learning classification techniques have been successfully used while building data-driven intelligent predictive systems in various application areas including smartphone apps. For an effective context-aware system, context pre-modeling is considered as a key issue and task, as the representation of contextual data directly influences the predictive models. This paper mainly explores the role of major context pre-modeling tasks, such as context vectorization by defining a good numerical measu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
20
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

4
3

Authors

Journals

citations
Cited by 27 publications
(20 citation statements)
references
References 45 publications
0
20
0
Order By: Relevance
“…A right and optimal subset of the selected features in a problem domain is capable to minimize the overfitting problem through simplifying and generalizing the model as well as increases the model's accuracy [97]. Thus, "feature selection" [66,99] is considered as one of the primary concepts in machine learning that greatly affects the effectiveness and efficiency of the target machine learning model. Chi-squared test, Analysis of variance (ANOVA) test, Pearson's correlation coefficient, recursive feature elimination, are some popular techniques that can be used for feature selection.…”
Section: Dimensionality Reduction and Feature Learningmentioning
confidence: 99%
See 2 more Smart Citations
“…A right and optimal subset of the selected features in a problem domain is capable to minimize the overfitting problem through simplifying and generalizing the model as well as increases the model's accuracy [97]. Thus, "feature selection" [66,99] is considered as one of the primary concepts in machine learning that greatly affects the effectiveness and efficiency of the target machine learning model. Chi-squared test, Analysis of variance (ANOVA) test, Pearson's correlation coefficient, recursive feature elimination, are some popular techniques that can be used for feature selection.…”
Section: Dimensionality Reduction and Feature Learningmentioning
confidence: 99%
“…-Feature extraction: In a machine learning-based model or system, feature extraction techniques usually provide a better understanding of the data, a way to improve prediction accuracy, and to reduce computational cost or training time. The aim of "feature extraction" [66,99] is to reduce the number of features in a dataset by generating new ones from the existing ones and then discarding the original features. The majority of the information found in the original set of features can then be summarized using this new reduced set of features.…”
Section: Dimensionality Reduction and Feature Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…Before the data is ready for modeling, it’s necessary to use data summarization and visualization to audit the quality of the data and provide the information needed to process it. To ensure the quality of the data, the data pre-processing technique, which is typically the process of cleaning and transforming raw data [ 107 ] before processing and analysis is important. It also involves reformatting information, making data corrections, and merging data sets to enrich data.…”
Section: Understanding Data Science Modelingmentioning
confidence: 99%
“…In this paper, we propose a model based on the Isolation Forest algorithm. At first, we perform necessary preprocessing steps like categorical feature encoding, feature scaling [11] to extract fifteen essential features to fit into the proposed model. Finally, we applied five popular classification algorithms [14] such as Logistic Regression (LR), Support Vector Machine (SVM), AdaBoost Classifier (ABC), Naive Bayes (NB), and K-Nearest Neighbor (KNN) to evaluate the performance of our system.…”
Section: Introductionmentioning
confidence: 99%