1994
DOI: 10.1006/jpdc.1994.1102
|View full text |Cite
|
Sign up to set email alerts
|

Performance and Optimization of Data Prefetching Strategies in Scalable Multiprocessors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

1994
1994
1997
1997

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(9 citation statements)
references
References 0 publications
0
9
0
Order By: Relevance
“…If this additional bandwidth is not available, prefetching may even be counter-productive because it overloads the network. (This has been pointed out in earlier publications, e.g., see [20].) In the full-prefetching strategy, overloading of the network is increased because all threads initiate the transfers right after receiving the invocation message sent by the caller of the operation.…”
Section: Successive Over-relaxationmentioning
confidence: 89%
“…If this additional bandwidth is not available, prefetching may even be counter-productive because it overloads the network. (This has been pointed out in earlier publications, e.g., see [20].) In the full-prefetching strategy, overloading of the network is increased because all threads initiate the transfers right after receiving the invocation message sent by the caller of the operation.…”
Section: Successive Over-relaxationmentioning
confidence: 89%
“…The remote memory access latency could be modified to reflect better the latency variations of a specific interconnection network. Previous studies reported in [4] and [26] have shown that network contention reduces the overall gains from supporting multiple outstanding requests in the network. Hence we expect slight drops in performance of all memory consistency models once these negative effects are included.…”
Section: ' Imentioning
confidence: 99%
“…Relaxed memory consistency allows reordering of memory events and buffering or pipelining of remote memory accesses [l], [13], [15], [16], [29], [8]. Data prefetching attempts to hide long read latency by issuing read requests well ahead of time, with the expectation that the data will be available in the cache when it is referenced [21], [26]. Multithreading attempts to hide the long latency by context switching between several active threads, thus allowing the processor to perform useful work while waiting for remote requests or synchronization faults to complete [2], [3], [4], [241, [251, [281. Most of these studies are based on simulation results.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Characterizations of the distance between the data movement initiation and the use of the data have also been based on program averages, like the cache miss ratio used by [1] to model coarse grain multithreading and the prefetch distance used by [10] to model the effectiveness of prefetching techniques. The write-run metric proposed by [5] is based on the average number of writes by a processor to a shared data item before an access by another processor.…”
Section: Introductionmentioning
confidence: 99%