2016
DOI: 10.1007/978-3-319-49956-7_4
|View full text |Cite
|
Sign up to set email alerts
|

Formalizing Data Locality in Task Parallel Applications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
3
3

Relationship

2
4

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 18 publications
0
6
0
Order By: Relevance
“…Several recent works have focused on CRD profiles for predicting the performance of shared cache [21,33,76,79,89]. Recently, researchers attempted to use analytical model and sampling to speed up the performance prediction [13,47,68,70,71].…”
Section: Reuse Distance Analysis On Multicore Processorsmentioning
confidence: 99%
“…Several recent works have focused on CRD profiles for predicting the performance of shared cache [21,33,76,79,89]. Recently, researchers attempted to use analytical model and sampling to speed up the performance prediction [13,47,68,70,71].…”
Section: Reuse Distance Analysis On Multicore Processorsmentioning
confidence: 99%
“…Traditional graphics scheduling approaches have focused on keeping hardware resources busy [10]. Yet scheduling can have a drastic impact in the data locality properties of the applications as well [11], [12], [13]. Scheduling tasks that share data together can reduce bandwidth and improve performance (frame rate) as they will be able to keep reused data in smaller caches.…”
Section: I I S C H E D U L I N Gmentioning
confidence: 99%
“…We start with an example that shows how the overall performance of an application changes when executing with different schedules due to an increase in last-level cache misses (Section 2). We then propose a profiling tool and TaskInsight's data classification technique that allows to clearly di↵erentiate the schedules in terms of their data reuse patterns, using a data reuse graph as in [2] (Section 3). Later, we show how to connect this classification to changes in data reuse, changes in cache misses and changes in performance during the execution: first from the perspective of the private caches (temporal locality on a single-threaded execution, Section 4) and later from the shared caches (spatial locality on multi-threaded run, Section 5).…”
Section: A New Technique To Analyze Schedulers Based On Thementioning
confidence: 99%
“…At the same time, TaskInsight builds the application's (schedule-independent) reuse graph introduced in [2], and combines it with the co-running task sequence to compute the set of memory addresses used by each co-running set. This allows to model the sequence of co-running tasks over time, and use the analysis in Section 4 to analyze how much data was reused over time, but in the shared cache.…”
Section: Locality Of Shared Cachesmentioning
confidence: 99%
See 1 more Smart Citation