2019
DOI: 10.1109/tvcg.2019.2934335
|View full text |Cite
|
Sign up to set email alerts
|

Void-and-Cluster Sampling of Large Scattered Data and Trajectories

Abstract: We propose a data reduction technique for scattered data based on statistical sampling. Our void-and-cluster sampling technique finds a representative subset that is optimally distributed in the spatial domain with respect to the blue noise property. In addition, it can adapt to a given density function, which we use to sample regions of high complexity in the multivariate value domain more densely. Moreover, our sampling technique implicitly defines an ordering on the samples that enables progressive data loa… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
19
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(19 citation statements)
references
References 61 publications
0
19
0
Order By: Relevance
“…By exploiting task-dependent cost functions, they are able to obtain satisfying scatter plot designs, which compete with those crafted by humans. Yuan et al[YXX * 20] and Rapp et al [RPD20] also address the problem of overdraw in scatter plots where they select a subset of data points from a large data set such that the resulting patterns follows the density, yet has blue noise. Our work does not select subsets, but shows all data points, and introduce an additional, non-encoding dimension so that dots can become blue noise in the first place.…”
Section: Previous Workmentioning
confidence: 99%
“…By exploiting task-dependent cost functions, they are able to obtain satisfying scatter plot designs, which compete with those crafted by humans. Yuan et al[YXX * 20] and Rapp et al [RPD20] also address the problem of overdraw in scatter plots where they select a subset of data points from a large data set such that the resulting patterns follows the density, yet has blue noise. Our work does not select subsets, but shows all data points, and introduce an additional, non-encoding dimension so that dots can become blue noise in the first place.…”
Section: Previous Workmentioning
confidence: 99%
“…The main hurdle to overcome is the estimation of the PDF of the dataset in phase-space, which is ultimately used to compute the acceptance probability. Different solutions have been proposed, including kernel methods [30] and binning [12]. Although the binning strategy is scalable with respect to the number of data points, the quality of the estimation may be dependent on the number of bins used [31].…”
Section: Probability Map Estimationmentioning
confidence: 99%
“…For the visualization of scatter plots, a perceptual loss was proposed to assess whether the scatter plot of downselected data was consistent with the original dataset [29]. Methods similar to the one proposed here have been proposed in the past [12,30], but either do not scale with the number of dimensions (D) or instances (N). Rapp et al [30] proposed a method to display a multidimensional scatter plot by transforming the distribution into an arbitrary one.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In situ processing is typically performed with a limited computational budget and in a distributed memory environment. Initial studies in this space have used SPSS techniques to represent and store steady state vector field data in the form of streamlines [HTZ*19] and unsteady state vector field data in the form of pathlines [ACG*14, BJ15, COJ15, SBC18, SCB19, RPD19]. These methods use SPSS techniques to perform an intelligent sampling and reduction of large vector fields such that they can be accurately reconstructed and explored post hoc.…”
Section: Research Challengesmentioning
confidence: 99%