25 Years of the International Symposia on Computer Architecture (Selected Papers) 1998
DOI: 10.1145/285930.285991
|View full text |Cite
|
Sign up to set email alerts
|

Memory access buffering in multiprocessors

Abstract: In highly-pipelined machines, instructions and data are prefetched and buffered in both the processor and the cache. This is done to reduce the average memory access latency and to take advantage of memory interleaving.Lockup tree caches are designed to avoid processor blocking on a cache miss. Write buffers are often included in a pipelined machine to avoid processor waiting on writes. In a shared memory multiprocessor, there are more advantages in buffering memory requests, since each memory access has to tr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
42
0

Year Published

2012
2012
2020
2020

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 50 publications
(42 citation statements)
references
References 6 publications
0
42
0
Order By: Relevance
“…On the other hand, if the consistency model is relaxed, i.e. not all possible orderings between memory operations are enforced, propagation of unordered memory operations can be delayed until an order can be re-established through synchronization boundaries [15,25,37]. In other words, lazy coherence protocols exploit the fact that relaxed consistency models require memory to be consistent only at synchronization boundaries.…”
Section: Eager Versus Lazy Coherencementioning
confidence: 99%
See 1 more Smart Citation
“…On the other hand, if the consistency model is relaxed, i.e. not all possible orderings between memory operations are enforced, propagation of unordered memory operations can be delayed until an order can be re-established through synchronization boundaries [15,25,37]. In other words, lazy coherence protocols exploit the fact that relaxed consistency models require memory to be consistent only at synchronization boundaries.…”
Section: Eager Versus Lazy Coherencementioning
confidence: 99%
“…However, it is possible to degrade performance for infrequently written but frequently read lines, suggested by our implementation of CC-shared-to-L2. Coherence for relaxed consistency: Dubois and Scheurich [15,37] first gave insight into reducing coherence overhead in relaxed consistency models, particularly that the requirement of "coherence on synchronization points" is sufficient. Instead of enforcing coherence at every write (also referred as the SWMR property [41]), recent works [7,12,17,21,28,35,42] enforce coherence at synchronization boundaries by self-invalidating shared data in private caches.…”
Section: Related Workmentioning
confidence: 99%
“…Memory consistency models describe the rules that guarantee memory accesses will be predictable. There are several memory consistency models that have been proposed, including sequential consistency (SC) [17], weak consistency (WC) [18], processor consistency (PC) [19], release consistency (RC) [15], entry consistency (EC) [6], scope consistency (ScC) [14].…”
Section: Related Workmentioning
confidence: 99%
“…The Java shared-memory model is not the sequentially consistent model [13], which is commonly used for writing multi-threaded programs. Instead, Java employs a form of weak consistency [14], to allow for shared-data optimization opportunities. In the sequential consistency model, any update to a shared variable must be visible to all other threads.…”
Section: The Java Parallel Programming Modelmentioning
confidence: 99%