2020
DOI: 10.14778/3380750.3380758
|View full text |Cite
|
Sign up to set email alerts
|

Data-parallel query processing on non-uniform data

Abstract: Graphics processing units (GPUs) promise spectacular performance advantages when used as database coprocessors. Their massive compute capacity, however, is often hampered by control flow divergence caused by non-uniform data distributions. When data-parallel work items demand for different amounts or types of processing, instructions execute with lowered efficiency. Query compilation techniques---a recent advance in GPU-accelerated database processing---suffer fr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 22 publications
(9 citation statements)
references
References 24 publications
0
9
0
Order By: Relevance
“…Sitaridi et al [157] proposed the splitting of string matching into multiple steps to minimize divergence. More recently, DogQC [6] detailed the high cost of importing data using dictionary encoding for strings and instead proposed the use of techniques like push-down parallelism and lane refill to minimize the divergence encountered by GPU threads when processing variable-length string data.…”
Section: Selection/filtermentioning
confidence: 99%
See 4 more Smart Citations
“…Sitaridi et al [157] proposed the splitting of string matching into multiple steps to minimize divergence. More recently, DogQC [6] detailed the high cost of importing data using dictionary encoding for strings and instead proposed the use of techniques like push-down parallelism and lane refill to minimize the divergence encountered by GPU threads when processing variable-length string data.…”
Section: Selection/filtermentioning
confidence: 99%
“…The new generation of compiled database systems for GPUs have helped minimize the synchronization overhead encountered by pipelined systems [3,5,6]. This is because multiple operators are co-located within the same kernel for compiled systems, allowing them to synchronize between operators using lower overhead thread-block level synchronization intrinsics for operators within the same pipeline.…”
Section: Concurrency Controlmentioning
confidence: 99%
See 3 more Smart Citations