2016
DOI: 10.1016/j.ascom.2016.06.002
|View full text |Cite
|
Sign up to set email alerts
|

A communication efficient and scalable distributed data mining for the astronomical data

Abstract: Please cite this article as: Govada, A., Sahay, S.K., A communication efficient and scalable distributed data mining for the astronomical data. Astronomy and Computing (2016), http://dx. AbstractIn 2020, ∼ 60PB of archived data will be accessible to the astronomers. But to analyze such a paramount data will be a challenging task. This is basically due to the computational model used to download the data from complex geographically distributed archives to a central site and then analyzing it in the local system… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 13 publications
0
1
0
Order By: Relevance
“…PCA is particularly advantageous when dealing with datasets containing numerous features. Furthermore, a distributed load balancing PCA (DLPCA) extension has been proposed [53] to reduce user transmission and download costs, further enhancing PCA's utility in astronomy data analysis. The steps for implementing PCA include centering input data, calculating the covariance matrix, identifying eigenvalues and eigenvectors, selecting relevant eigenvectors, and generating the final reduced-dimension data.…”
Section: Unsupervised Techniquesmentioning
confidence: 99%
“…PCA is particularly advantageous when dealing with datasets containing numerous features. Furthermore, a distributed load balancing PCA (DLPCA) extension has been proposed [53] to reduce user transmission and download costs, further enhancing PCA's utility in astronomy data analysis. The steps for implementing PCA include centering input data, calculating the covariance matrix, identifying eigenvalues and eigenvectors, selecting relevant eigenvectors, and generating the final reduced-dimension data.…”
Section: Unsupervised Techniquesmentioning
confidence: 99%