Proceedings of the 2020 ACM SIGMOD International Conference on Management of Data 2020
DOI: 10.1145/3318464.3389732
|View full text |Cite
|
Sign up to set email alerts
|

Database Benchmarking for Supporting Real-Time Interactive Querying of Large Data

Abstract: In this paper, we present a new benchmark to validate the suitability of database systems for interactive visualization workloads. While there exist proposals for evaluating database systems on interactive data exploration workloads, none rely on real user traces for database benchmarking. To this end, our long term goal is to collect user traces that represent workloads with different exploration characteristics. In this paper, we present an initial benchmark that focuses on "crossfilter"-style applications, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4
1

Relationship

2
7

Authors

Journals

citations
Cited by 26 publications
(16 citation statements)
references
References 58 publications
0
15
0
1
Order By: Relevance
“…We designed four visual analytics tasks (see Table 2) for each dataset based on prior studies of data analysis [2,3,36,37]. These four tasks cover all three analysis task classes discussed by Battle et al [2]: quantitative, qualitative, and exploratory. T1 and T2 are focused tasks; T1 involves two data attributes, while T2 involves three data attributes.…”
Section: Methodsmentioning
confidence: 99%
“…We designed four visual analytics tasks (see Table 2) for each dataset based on prior studies of data analysis [2,3,36,37]. These four tasks cover all three analysis task classes discussed by Battle et al [2]: quantitative, qualitative, and exploratory. T1 and T2 are focused tasks; T1 involves two data attributes, while T2 involves three data attributes.…”
Section: Methodsmentioning
confidence: 99%
“…Each trace is 3 minutes long, with 20ms average think time. For Falcon, we used the 70 traces from [7]. The interface used to collect these traces differs from the interface in the Falcon paper [53] by one chart (a bar chart instead of the heat map in [53]).…”
Section: Methodsmentioning
confidence: 99%
“…(4) Prefetching and re-partitioning: Interactions impose an even stricter latency requirement for visualizations [1]. Based on the idea of partial execution in subsection 2.1, partially processed data can be brought back to the client earlier so that a downstream interaction parameterized by such data will trigger a fast partial execution.…”
Section: Middleware Optimization Dynamicmentioning
confidence: 99%
“…As a result, the visualizations remain lightweight, stand-alone and agnostic of the optimization work behind. Second, Vega is also expressive enough to capture the computational complexity of most visualization interfaces, including those tested in recent database management system (DBMS) benchmarks designed for visual exploration scenarios [1,4]. Third, Vega is the backbone of a popular ecosystem of visualization tools, including Vega-Lite [9], Voyager [12], and Falcon [6], so making improvements to Vega is of interest to thousands of data enthusiasts, researchers, and companies worldwide.…”
Section: Introductionmentioning
confidence: 99%