2023
DOI: 10.15294/sji.v10i2.44027
|View full text |Cite
|
Sign up to set email alerts
|

Classification of Spiral and Non-Spiral Galaxies using Decision Tree Analysis and Random Forest Model: A Study on the Zoo Galaxy Dataset

Abstract: Purpose: The goal of this research is to create a precise prediction model that can differentiate between spiral and non-spiral galaxies using the Zoo galaxy dataset. Decision tree analysis and random forest models will be used to construct the model, and various conditions within the dataset will be employed to classify the data accurately. The model's performance will be evaluated using a confusion matrix, and the probability of predicting spiral galaxies will be analyzed. The research will also investigate … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 24 publications
0
2
0
Order By: Relevance
“…This technique addresses the constraints encountered with a single Decision Tree, including issues such as overfitting and potential bias of the dataset. RF enhances predictive power by amalgamating insights from multiple decision trees, resulting in robust and precise predictions [34], [35], [36]. The RF algorithm begins with random sample selection, decision-tree construction, and voting or averaging to generate a final prediction.…”
Section: Random Forest (Rf)mentioning
confidence: 99%
“…This technique addresses the constraints encountered with a single Decision Tree, including issues such as overfitting and potential bias of the dataset. RF enhances predictive power by amalgamating insights from multiple decision trees, resulting in robust and precise predictions [34], [35], [36]. The RF algorithm begins with random sample selection, decision-tree construction, and voting or averaging to generate a final prediction.…”
Section: Random Forest (Rf)mentioning
confidence: 99%
“…The entropy equation serves as a pivotal tool in DT analysis, particularly when calculating the impurity at a node, represented as: πΈπ‘›π‘‘π‘Ÿπ‘œπ‘π‘¦ (𝑃) = βˆ’ βˆ‘ 𝑝(𝑖). log 2 (𝑝(𝑖)) (8) In this equation, πΈπ‘›π‘‘π‘Ÿπ‘œπ‘π‘¦ (𝑃) represents the entropy of the dataset P, where 𝑝(𝑖) denotes the probability that an instance in dataset P belongs to class i [67].…”
Section: Decision Treementioning
confidence: 99%