2014 Second International Symposium on Computing and Networking 2014
DOI: 10.1109/candar.2014.49
|View full text |Cite
|
Sign up to set email alerts
|

Hinting for Auto-Memoization Processor Based on Static Binary Analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2018
2018
2018
2018

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 5 publications
0
5
0
Order By: Relevance
“…DTM is a reuse technique that operates on traces of instructions and is often implemented on top of Von Neumann‐based superscalar architectures, with further studies that include speculative execution . Speculative execution often improves the reuse rate of traces, because it enables reuse based on speculative values for input operands.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…DTM is a reuse technique that operates on traces of instructions and is often implemented on top of Von Neumann‐based superscalar architectures, with further studies that include speculative execution . Speculative execution often improves the reuse rate of traces, because it enables reuse based on speculative values for input operands.…”
Section: Related Workmentioning
confidence: 99%
“…15 The size of each operation, ie, the reuse granularity, can vary from a single instruction 16 to groups of instructions, such as functions, 11 expressions, 17 basic blocks, 18 sub-blocks, 19 or traces. 20 DTM 10 is a reuse technique that operates on traces of instructions and is often implemented on top of Von Neumann-based superscalar architectures, [16][17][18][19][20][21][22][23][24] with further studies that include speculative execution. [25][26][27][28] Speculative execution often improves the reuse rate of traces, because it enables reuse based on speculative values for input operands.…”
Section: Related Workmentioning
confidence: 99%
“…22 Moreover, other studies have implemented similar memoization schemes into ARM-based superscalar processors. [23][24][25][26] Some works have also explored the reuse of computation in the GPU domain. 27 For instance, redundant fragment shader executions have been reused on a mobile GPU through hardware memoization.…”
Section: Related Workmentioning
confidence: 99%
“…Since it is difficult to track these global changes at runtime, existing hardware memoization approaches apply to CHAPTER 1. INTRODUCTION single instructions or blocks of instructions [139,64,66,82,49,50,91,145], whereas function level memoization has only been exploited in software based solutions [132,84,110,168]. Our memoization scheme is different as it is function level and hardware based, since our work is focused on GPUs and graphical applications where it is easier to track changes to global data and no mutable state or side-effects exist.…”
Section: Eliminating Redundant Fragment Shader Executionsmentioning
confidence: 99%
“…As pointed out in [134], this concept is important because computation re-use is lucrative only when the cost of accessing the structures used for memoization is smaller than the benefit of skipping the actual computation. For this reason prior work on memoization either tries to perform memoization for multiple instructions [66,82,49,50,91,145] or for long latency operations [64].…”
Section: Task-level Complexitymentioning
confidence: 99%