1998
DOI: 10.1109/12.689650
|View full text |Cite
|
Sign up to set email alerts
|

Combining trace sampling with single pass methods for efficient cache simulation

Abstract: The design of the memory hierarchy is crucial to the performance of high performance computer systems. The incorporation of multiple levels of caches into the memory hierarchy is known to increase the performance of high end machines, but the development of architectural prototypes of various memory hierarchy designs is costly and time consuming. In this paper, we will describe a single pass method used in combination with trace sampling techniques to produce a fast and accurate approach for simulating multipl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
34
0

Year Published

2003
2003
2019
2019

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 52 publications
(37 citation statements)
references
References 12 publications
0
34
0
Order By: Relevance
“…Much of the accelerated simulator literature falls under this category, only differing in which portions of the code are selected and how to fastforward to those specific points in the program [12,19,27]. Several techniques have been proposed to warmup cache state prior to measurement in order to obtain more accurate simulation results [1,4,5,10,11,27]. The MTR adds multiprocessor and directory support to these techniques.…”
Section: Related Workmentioning
confidence: 99%
“…Much of the accelerated simulator literature falls under this category, only differing in which portions of the code are selected and how to fastforward to those specific points in the program [12,19,27]. Several techniques have been proposed to warmup cache state prior to measurement in order to obtain more accurate simulation results [1,4,5,10,11,27]. The MTR adds multiprocessor and directory support to these techniques.…”
Section: Related Workmentioning
confidence: 99%
“…Armed with equation (6), we can compute the expected stack distance distribution of an application given its RDS as follows: First, we compute the expected stack distance of each distinct reuse distance in the RDS using (6). Then, by weighting each of the expected stack distances with the frequency of the corresponding reuse distance in the RDS we get the expected stack distance distribution.…”
Section: Cache Modelmentioning
confidence: 99%
“…In fact, most of the prior research on the cold-start problem has been done on cache warmup. Various approaches have been proposed such as no warmup, stale state (also called stitch) [13], fixed warmup [1], cache miss rate estimators [14], no-state-loss [12,15], minimal subset evaluation (MSE) [16], memory reference reuse latency (MRRL) [17], boundary line reuse latency (BLRL) [8,18], self-monitored adaptive cache warmup (SMA) [19], memory hierarchy state (MHS) [5], memory timestamp record (MRT) [7], etc.…”
Section: Cache Warmupmentioning
confidence: 99%