2016 XLII Latin American Computing Conference (CLEI) 2016
DOI: 10.1109/clei.2016.7833394
|View full text |Cite
|
Sign up to set email alerts
|

Potential benefits of a block-space GPU approach for discrete tetrahedral domains

Abstract: The study of data-parallel domain re-organization and thread-mapping techniques are relevant topics as they can increase the efficiency of GPU computations when working on spatial discrete domains with non-box-shaped geometry. In this work we study the potential benefits of applying a succint data re-organization of a tetrahedral data-parallel domain of size O(n 3 ) combined with an efficient block-space GPU map of the form g(λ) : N → N 3 . Results from the analysis suggest that in theory the combination of th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
3
3

Relationship

2
4

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 12 publications
0
4
0
Order By: Relevance
“…Navarro et al [11,12,13] proposed a GPU block-space mapping for 2 and 3-simplex domains based on a linear numbering of discrete elements. Authors report an empirical speedup of up to 1.5× and 2.3× over a bounding-box approach, for 2 and 3-simplices, respectively.…”
Section: Gpu Processing In Complex Domainsmentioning
confidence: 99%
See 1 more Smart Citation
“…Navarro et al [11,12,13] proposed a GPU block-space mapping for 2 and 3-simplex domains based on a linear numbering of discrete elements. Authors report an empirical speedup of up to 1.5× and 2.3× over a bounding-box approach, for 2 and 3-simplices, respectively.…”
Section: Gpu Processing In Complex Domainsmentioning
confidence: 99%
“…Thanks to the 40GB of GPU memory of the A100, it was possible to push the maximum problem size up to 2 16 , except for the curve of S 1x1 that could not reach the maximum size because of CUDA's grid size limits. The TITAN RTX under-performed in comparison with the other GPUs for n ≤ 2 12 . Past that value, speedup increases above 1 for all values of ρ, reaching a top speedup of ∼ 3.2×.…”
Section: Performance Plotsmentioning
confidence: 99%
“…Ries et al [15] developed in 2009 a method to compute the inverse of triangular matrices by developing a recursive parallel space mapping from a compact rectangular domain using GPU. Navarro et al proposed in 2014 a GPU block-space mapping for 2-simplex and 3-simplex shaped data [9,8,10]. Navarro et al [7,11] expanded the idea of GPU thread mapping for fractal domains, by proposing the λ(ω) map for NBB fractals.…”
Section: Related Workmentioning
confidence: 99%
“…Here, n ∈ N is the linear size of the fractal along one axis, k ∈ N the number of self-similar replicas to generate for the next scale level and s ∈ N the growth ratio of n in the next scale level, along an axis. For example, the Sierpiński Carpet (Figure 1) is F 8,3 n and the Sierpiński Triangle (Figure 2) is F 3,2 n . Many different NBB fractals can be described using the same parameters.…”
mentioning
confidence: 99%