2014
DOI: 10.2478/amcs-2014-0011
|View full text |Cite
|
Sign up to set email alerts
|

An algorithm for reducing the dimension and size of a sample for data exploration procedures

Abstract: The paper deals with the issue of reducing the dimension and size of a data set (random sample) for exploratory data analysis procedures. The concept of the algorithm investigated here is based on linear transformation to a space of a smaller dimension, while retaining as much as possible the same distances between particular elements. Elements of the transformation matrix are computed using the metaheuristics of parallel fast simulated annealing. Moreover, elimination of or a decrease in importance is perform… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
5
3
1

Relationship

3
6

Authors

Journals

citations
Cited by 19 publications
(5 citation statements)
references
References 47 publications
0
5
0
Order By: Relevance
“…It is a case for traditional Principal Component Analysis algorithm. Here, we consider similar strategy assuring preserving distances between data sample elements and the metaheuristics of a Parallel Fast Simulated Annealing [16,17,22]. This approach is applicable to all EDA procedures.…”
Section: Optimization Inspired By Nature As a Tool For Eda -A Survey mentioning
confidence: 99%
“…It is a case for traditional Principal Component Analysis algorithm. Here, we consider similar strategy assuring preserving distances between data sample elements and the metaheuristics of a Parallel Fast Simulated Annealing [16,17,22]. This approach is applicable to all EDA procedures.…”
Section: Optimization Inspired By Nature As a Tool For Eda -A Survey mentioning
confidence: 99%
“…A detailed description of the methodology presented here can be found in the work [40] as well as in the paper [41] which will appear soon.…”
Section: Summary and Finals Remarksmentioning
confidence: 99%
“…The latter arises mostly from a number of phenomena occurring in data sets of this type, known in literature as "the curse of multidimensionality". Above all, this includes the exponential growth in sample size, necessary to achieve appropriate effectiveness of data analysis methods with increasing dimension, as well as the vanishing difference between near and far points (norm concentration) using standard distance metrics [30].…”
Section: Data Volume Problem In Observational Astronomymentioning
confidence: 99%