2013 International Conference on Field-Programmable Technology (FPT) 2013
DOI: 10.1109/fpt.2013.6718396
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive compression for instruction code of Coarse Grained Reconfigurable Architectures

Abstract: Coarse Grained Reconfigurable Architecture (CGRA) achieves high performance by exploiting instructionlevel parallelism with software pipeline. Large instruction memory is, however, a critical problem of CGRA, which requires large silicon area and power consumption. Code compression is a promising technique to reduce the memory area, bandwidth requirements, and power consumption. We present an adaptive code compression scheme for CGRA instructions based on dictionary-based compression, where compression mode an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 7 publications
0
5
0
Order By: Relevance
“…Aslam et al [31] uses state-of-the-art dictionary methods and reorganized the PEs to improve the compression in the dictionary. The approach taken by Chung et al [32,33] exploits the spatial and temporal redundancy from the configuration stream and saves the most frequently occurring values in a dictionary. The latency of their decompressor is two cycles and can be pipelined, but the overhead of the decompressor is not shown in their papers.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Aslam et al [31] uses state-of-the-art dictionary methods and reorganized the PEs to improve the compression in the dictionary. The approach taken by Chung et al [32,33] exploits the spatial and temporal redundancy from the configuration stream and saves the most frequently occurring values in a dictionary. The latency of their decompressor is two cycles and can be pipelined, but the overhead of the decompressor is not shown in their papers.…”
Section: Related Workmentioning
confidence: 99%
“…For example, in Figure 2, HReA [18] uses GCM 1024 bits wide and 128 lines deep feeding the context to PEs and it takes 42% of the entire chip area and 38% of the chip power consumption. To reduce context-fetching overhead and context-memory footprint, most existing context-reduction frameworks [31][32][33][34][35][36] rely on the statistical analysis over the context bitstream in the pre-silicon phase, and their compressed context is encoded after the original context-generation phase. These approaches can be classified as post-contextgeneration method [37].…”
Section: Introductionmentioning
confidence: 99%
“…In [27], they proposed a new method to reduce the energy of the context switching process by decreasing the number of active lines of the context memory and also reducing the number of transition bits on the CSN. In [28], they used the differential loading technique to reduce the number of transition bits on the CSN. But, Kim and Mahapatra [29,30] neither stored nor transmitted the unnecessary bits of the configuration words to reduce the energy of the CGRA.…”
Section: Related Workmentioning
confidence: 99%
“…By fragmenting a BB into multiple contexts, one branching instruction (per context) will be added to the Ni. On the other hand, based on the efficiency of the CGRA compiler, some PEs in the PEN cannot be mapped and they will be idle [28]. Thus, we have added a new parameter for that effect, called α.…”
Section: Figure 2 Input High-level Programmentioning
confidence: 99%
“…Due to the complexity of the compression, a genetic algorithm and an integer linear program (ILP) are used. A dictionary-based compression is proposed in [10]. This method separates two types of dictionaries based on the locality of a mapped application.…”
Section: Related Workmentioning
confidence: 99%