2014
DOI: 10.1504/ijes.2014.065000
|View full text |Cite
|
Sign up to set email alerts
|

Insights on memory controller scaling in multi-core embedded systems

Abstract: Abstract:In recent years, the growth on the number of cores as well as the frequency of cores along different processor generations has proportionally increased bandwidth needs simultaneously in both CPU and GPU systems. In order to address the communication latency between CPU and GPU memories in recent implementation of heterogeneous mobile embedded systems with hard or firm real-time requirements, sharing the same address space adds significant levels of contention. In addition, when heterogeneous cores are… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
19
0

Year Published

2015
2015
2018
2018

Publication Types

Select...
7

Relationship

2
5

Authors

Journals

citations
Cited by 9 publications
(19 citation statements)
references
References 16 publications
0
19
0
Order By: Relevance
“…• Revisiting the operating system (OS) concept of address space used by Marino and Li's report [15], in RAM ON the novel concept of region is defined as an address space range dedicated to different sets of cores (CPU, GPU, or both), caches and respective interconnection. The inclusion of the two latter elements differentiate RAM ON from Non Uniform Memory Access (NUMAnode) mechanism in Linux OS.…”
mentioning
confidence: 99%
See 2 more Smart Citations
“…• Revisiting the operating system (OS) concept of address space used by Marino and Li's report [15], in RAM ON the novel concept of region is defined as an address space range dedicated to different sets of cores (CPU, GPU, or both), caches and respective interconnection. The inclusion of the two latter elements differentiate RAM ON from Non Uniform Memory Access (NUMAnode) mechanism in Linux OS.…”
mentioning
confidence: 99%
“…• An evaluation on system implications for larger numbers of MCs in heterogeneous regions via using opticaland RF-based interfaces (signal modulation) rather than traditional digital transmission (where, to transmit a "0" or a "1", the whole line should be entirely set to the respective level) developed in [15].…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…In addition, methods that combine DVFS [22] to shallower transaction queues under different memory traffic intensities and applications are likely to be considered. Furthermore, heterogeneous systems [30] offer interesting opportunities to evaluate transaction queue size reduction.…”
Section: Conclusion and Future Plansmentioning
confidence: 99%
“…For example, assembly of human genome using PASHA software took around 21 hours on a 8-core workstation with 72 GB memory (Liu et al, 2011). Hardware accelerators like GPU and FPGAs are used along with processors to reduce this execution time by running the program on multiple computation units in parallel (Lin and Lin, 2014;Okuyama et al, 2012;Halstead et al, 2014;Marino and Li, 2014). Some bioinformatics applications along with assembly programs have been accelerated by FPGA-based accelerators.…”
Section: Introductionmentioning
confidence: 99%