2022
DOI: 10.29408/edumatic.v6i1.5613
|View full text |Cite
|
Sign up to set email alerts
|

Comparison of Naïve Bayes Algorithm and XGBoost on Local Product Review Text Classification

Abstract: Online reviews are critical in supporting purchasing decisions because, with the development of e-commerce, there are more and more fake reviews, so more and more consumers are worried about being deceived in online shopping. Sentiment analysis can be applied to Marketplace product reviews. This study aims to compare the two categories of Naïve Bayes and XGBoost by using the two vector spaces wod2vec and TFIDF. The methods used in this research are data collection, data cleaning, data labelling, data pre-proce… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
7
0
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 10 publications
0
7
0
1
Order By: Relevance
“…The XGBoost approach was found to be more effective in solving classification problems by Hendrawan et al (2022) in their study on e-commerce product reviews. Elmitwally (2020) claimed in his research that he used XGBoost to accomplish the classification problem’s highest prediction performance.…”
Section: Methodsmentioning
confidence: 99%
“…The XGBoost approach was found to be more effective in solving classification problems by Hendrawan et al (2022) in their study on e-commerce product reviews. Elmitwally (2020) claimed in his research that he used XGBoost to accomplish the classification problem’s highest prediction performance.…”
Section: Methodsmentioning
confidence: 99%
“…It is particularly suitable for situations where the data distribution is not explicitly known or may exhibit non-standard characteristics. XGBoost employs a boosting technique to improve model performance sequentially by correcting errors ( Hendrawan et al, 2022 ; Arif Ali et al, 2023 ). It has robustness in handling linear and non-linear relationships, including missing data ( Hendrawan et al, 2022 ; Arif Ali et al, 2023 ).…”
Section: Methodsmentioning
confidence: 99%
“…XGBoost employs a boosting technique to improve model performance sequentially by correcting errors ( Hendrawan et al, 2022 ; Arif Ali et al, 2023 ). It has robustness in handling linear and non-linear relationships, including missing data ( Hendrawan et al, 2022 ; Arif Ali et al, 2023 ). RF combines multiple decision trees to improve overall prediction accuracy and reduce overfitting ( Belgiu and Drăguţ, 2016 ).…”
Section: Methodsmentioning
confidence: 99%
“…Future Proofing: Even if the XGBRFClassifier does not significantly outperform Naïve Bayes on our current dataset, it may be more adaptable to future changes in data distribution or feature sets [116]. Naïve Bayes is relatively simple and may not handle data shifts or feature additions as gracefully as the XGBRFClassifier [117].…”
mentioning
confidence: 96%