2016 IEEE International Parallel and Distributed Processing Symposium (IPDPS) 2016
DOI: 10.1109/ipdps.2016.49
|View full text |Cite
|
Sign up to set email alerts
|

CATA: Criticality Aware Task Acceleration for Multicore Processors

Abstract: Abstract-Managing criticality in task-based programming models opens a wide range of performance and power optimization opportunities in future manycore systems. Criticality aware task schedulers can benefit from these opportunities by scheduling tasks to the most appropriate cores. However, these schedulers may suffer from priority inversion and static binding problems that limit their expected improvements.Based on the observation that task criticality information can be exploited to drive hardware reconfigu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
22
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
6
1
1

Relationship

5
3

Authors

Journals

citations
Cited by 19 publications
(22 citation statements)
references
References 51 publications
0
22
0
Order By: Relevance
“…Research on DVFS with parallel workloads has focused on using DVFS to accelerate the critical path in applications [8], [10] or to improve the overall energy efficiency of a system when running big data workloads [14]. In this work, we focus on parallel workloads in a widely-used runtime such as OpenMP and with several goals: reduce execution time, reduce EDP, and reduce power consumption.…”
Section: Dynamic Voltage and Frequency Scalingmentioning
confidence: 99%
“…Research on DVFS with parallel workloads has focused on using DVFS to accelerate the critical path in applications [8], [10] or to improve the overall energy efficiency of a system when running big data workloads [14]. In this work, we focus on parallel workloads in a widely-used runtime such as OpenMP and with several goals: reduce execution time, reduce EDP, and reduce power consumption.…”
Section: Dynamic Voltage and Frequency Scalingmentioning
confidence: 99%
“…The runtime system dynamically schedules tasks when all their inputs are ready and, when the execution of a task finishes, its outputs become ready for the next tasks. This model decouples the hardware from the application, enabling many optimizations at the runtime system level in a generic and application-agnostic way [2,10,11,25,30,41].…”
Section: Task-based Programming Modelsmentioning
confidence: 99%
“…With this information the runtime system manages the parallel execution following a data-flow scheme, scheduling tasks to cores and taking care of synchronization between tasks. Decoupling the application from the architecture eases programmability and allows to leverage the runtime system information to drive optimizations in a generic and application-agnostic way [2,10,11,25,30,41].…”
Section: Introductionmentioning
confidence: 99%
“…The input and output information allows the runtime system to transparently manage GPUs [47], [58], stacked DRAM memories [59], multi-node clusters [60], and scratchpad memories [61]. With some additional hardware support, the runtime system can do value approximation [62], software-guided prefetching [13], dead block prediction [16], accelerate critical tasks [63], reduce coherence traffic in CC-NUMA systems [64], [65], and optimise communications in producer-consumer task relationships [66].…”
Section: Task-based Programming Modelsmentioning
confidence: 99%