Proceedings of the 19th International Conference on Architectural Support for Programming Languages and Operating Systems 2014
DOI: 10.1145/2541940.2541965
|View full text |Cite
|
Sign up to set email alerts
|

Scale-out NUMA

Abstract: Emerging datacenter applications operate on vast datasets that are kept in DRAM to minimize latency. The large number of servers needed to accommodate this massive memory footprint requires frequent server-to-server communication in applications such as key-value stores and graph-based applications that rely on large irregular data structures. The fine-grained nature of the accesses is a poor match to commodity networking technologies, including RDMA, which incur delays of 10-1000x over local DRAM operations.W… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
54
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 89 publications
(54 citation statements)
references
References 50 publications
0
54
0
Order By: Relevance
“…10 describes the emulation platform for the so-NUMA architecture. This platform [47] is designed to (i) run server nodes at regular wall-clock speed, and (ii) approximate the latency and bandwidth of the fabric. The emulation platform relies on hardware virtualization to create a RackOut unit of up to 16 nodes.…”
Section: Experimental Methodologymentioning
confidence: 99%
See 3 more Smart Citations
“…10 describes the emulation platform for the so-NUMA architecture. This platform [47] is designed to (i) run server nodes at regular wall-clock speed, and (ii) approximate the latency and bandwidth of the fabric. The emulation platform relies on hardware virtualization to create a RackOut unit of up to 16 nodes.…”
Section: Experimental Methodologymentioning
confidence: 99%
“…EMC/Isilon [24]) solutions to clients connected via a conventional network. AppliedMicro's X-Gene2 server SoC [40] and Oracle's Sonoma [41] integrate the RDMA controller directly on chip, HP Moonshot [30] combines low-power processors with RDMA NICs, and research proposals further argue for on-chip support for one-sided remote access primitives [18,47]. Building larger logical entities using such rack-scale memory pooling approaches instead of the cache-coherent NUMA approach comes at a lower cost and complexity.…”
Section: Architectural Building Blocksmentioning
confidence: 99%
See 2 more Smart Citations
“…Emerging low-latency rack-scale fabrics [7] may provide a way to efficiently aggregate memory to reduce the pressure on dynamic replication.…”
Section: Dynamic Replicationmentioning
confidence: 99%