1992
DOI: 10.1145/146628.139705
|View full text |Cite
|
Sign up to set email alerts
|

Comparative performance evaluation of cache-coherent NUMA and COMA architectures

Abstract: Two interesting variations of large-scale shared-memory machines that have recently emerged are cache-coherent mmumform-memory-access machines (CC-NUMA) and cacheonly memory architectures (COMA). They both have distributed main memory and use directory-based cache coherence. Unlike CC-NUMA, however, COMA machines automatically migrate and replicate data at the main-memoty level in cache-line sized chunks. This paper compares the performance of these two classes of machines.We first present a qualitative model … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

1996
1996
2021
2021

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(1 citation statement)
references
References 7 publications
0
1
0
Order By: Relevance
“…Then, the modi cation of this shared variable results in invalidating the cache lines including the lock variable of waiting threads and even lock holders. In particular, when a cache line is shared among threads in a NUMA environment, frequent invocations of the cache line invalidation to maintain the cache coherence become a signi cant overhead [19], [29]. In the case of SRL(tree,i), when it dynamically allocates slvs, the cache line for atomic operations can be shared.…”
Section: B Performance Evaluation 1) Lock Acquisition Latencymentioning
confidence: 99%
“…Then, the modi cation of this shared variable results in invalidating the cache lines including the lock variable of waiting threads and even lock holders. In particular, when a cache line is shared among threads in a NUMA environment, frequent invocations of the cache line invalidation to maintain the cache coherence become a signi cant overhead [19], [29]. In the case of SRL(tree,i), when it dynamically allocates slvs, the cache line for atomic operations can be shared.…”
Section: B Performance Evaluation 1) Lock Acquisition Latencymentioning
confidence: 99%