Proceedings of the 2015 International Symposium on Memory Systems 2015
DOI: 10.1145/2818950.2818955
|View full text |Cite
|
Sign up to set email alerts
|

Another Trip to the Wall

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 39 publications
(9 citation statements)
references
References 10 publications
0
9
0
Order By: Relevance
“…Our results (and results from previous studies) suggest that a future ARCHER service could even benefit from architectures where HBM-like technologies with limited capacity replace main memory, rather than using a hybrid solution (such as the MC-DRAM+DRAM seen on the Intel Xeon Phi). The reasoning here is that using HBM technologies as a main memory replacement allows applications to access the best performance without application code modifications whereas in the hybrid approach the only way to use the HBM without code modification is as an additional, large cache level, which can limit the performance gains available [6]. Another option would be to use a combination of processors with high memory bandwidth alongside processors with high memory capacity.…”
Section: Discussionmentioning
confidence: 99%
“…Our results (and results from previous studies) suggest that a future ARCHER service could even benefit from architectures where HBM-like technologies with limited capacity replace main memory, rather than using a hybrid solution (such as the MC-DRAM+DRAM seen on the Intel Xeon Phi). The reasoning here is that using HBM technologies as a main memory replacement allows applications to access the best performance without application code modifications whereas in the hybrid approach the only way to use the HBM without code modification is as an additional, large cache level, which can limit the performance gains available [6]. Another option would be to use a combination of processors with high memory bandwidth alongside processors with high memory capacity.…”
Section: Discussionmentioning
confidence: 99%
“…To understand cloud workloads' memory bandwidth utilization trend, we show the fleet-wide memory bandwidth consumption increase over the three generations of servers in Figure 4. With the most recent generation (Gen 5), the 1-minute average memory bandwidth utilization shows that the majority of the fleet is While mirco-benchmarks can generally drive memory bandwidth utilization to >80%, we observe that production workloads rarely exceed 60% memory bandwidth utilization, as any further increase results in an exponential increase in memory latency [32]. Thus, we classify workloads with higher than 60% bandwidth utilization as memory bandwidth bound (shaded red in Figure 4).…”
Section: Memory Bandwidth Scaling Challengementioning
confidence: 95%
“…The main drawback of this technique is the decreased computing speed that leads to increased application run times. This issue is partially mitigated because several HPC applications and benchmarks are not CPU-bound but present a memory and I/O bottleneck (Marjanović et al, 2014;Radulovic et al, 2015;Zivanovic et al, 2017); reducing the frequency of the computing units used by these jobs does not impact severely their time-to-solution (TtS) (Auweter et al, 2014). While in the rest of the article, we will refer explicitly to frequency scaling, our conclusions can also be applied to systems using Intel's Running Average Power Limit (RAPL) (David et al, 2010), that does not directly change the computing nodes clock frequency but enforces a socket-level power cap.…”
Section: Introductionmentioning
confidence: 99%