Proceedings of the Seventh International Conference on Architectural Support for Programming Languages and Operating Systems 1996
DOI: 10.1145/237090.237151
|View full text |Cite
|
Sign up to set email alerts
|

Thread scheduling for cache locality

Abstract: This paper describes a method to improve the cache locality of sequential programs by scheduling fine-grained threads. The algorithm relies upon hints provided at the time of thread creation to determine a thread execution order likely to reduce cache misses. This technique may be particularly valuable when compiler-directed tiling is not feasible. Experiments with several application programs, on two systems with different cache structures, show that our thread scheduling method can improve program performanc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
37
0

Year Published

1997
1997
2014
2014

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 69 publications
(37 citation statements)
references
References 36 publications
0
37
0
Order By: Relevance
“…In [15], Philbin et al reordered loops in sequential applications to improve locality, using information about data accesses. In the realm of task-parallelism, Chen et al proposed scheduling concurrent execution in order to promote cache sharing on CMPs [2].…”
Section: Related Workmentioning
confidence: 99%
“…In [15], Philbin et al reordered loops in sequential applications to improve locality, using information about data accesses. In the realm of task-parallelism, Chen et al proposed scheduling concurrent execution in order to promote cache sharing on CMPs [2].…”
Section: Related Workmentioning
confidence: 99%
“…While their approach is not particularly well-suited to non-real- time systems, their micro-benchmark results do indicate that intelligent co-scheduling of cooperative threads can reduce the number of L2 misses substantially. Philbin et al [30] studied the possibility of reducing cache misses for sequential programs through intelligent scheduling of fine-grained threads. Their approach relies on memory access hints in the program to identify threads that should execute in close temporal proximity in order to promote cache re-use.…”
Section: Related Workmentioning
confidence: 99%
“…For instance, Philbin et al [11] formalise the problem of locality-aware thread scheduling for a single-core processor. In other work by Tam et al [14], threads are grouped based on data-locality for multi-threaded multi-core processors, introducing a metric of thread similarity.…”
Section: Related Workmentioning
confidence: 99%