2014
DOI: 10.1587/transfun.e97.a.1027
|View full text |Cite
|
Sign up to set email alerts
|

An Efficient Parallel SOVA-Based Turbo Decoder for Software Defined Radio on GPU

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 16 publications
0
5
0
Order By: Relevance
“…To create a socket, four parameters are required: its datatype (given as a template parameter), its associated task, its name, and its size. Finally, a "codelet" function need to be set (lines [14][15][16][17][18][19][20][21][22][23][24][25][26]. This codelet will be called when the task will be triggered.…”
Section: Elementary Componentsmentioning
confidence: 99%
See 1 more Smart Citation
“…To create a socket, four parameters are required: its datatype (given as a template parameter), its associated task, its name, and its size. Finally, a "codelet" function need to be set (lines [14][15][16][17][18][19][20][21][22][23][24][25][26]. This codelet will be called when the task will be triggered.…”
Section: Elementary Componentsmentioning
confidence: 99%
“…Many SDR elementary blocks have been optimized for Intel® and ARM® CPUs. High throughput results have been achieved on GPUs; 19‐23 latency results are is still too high however to meet real time constraints and to compete with CPU implementations 22,24‐33 . This is mainly due to data transfers between the host (CPUs) and the device (GPUs), and to the nature of GPU designs, which are not optimized for latency efficiency.…”
Section: Introductionmentioning
confidence: 99%
“…In [16], [20], multiple trellisstate computations are performed in parallel in the SIMD units. In [9]- [18], [20], the decoded frame is split into sub-blocks that are processed in parallel in the SIMD units. An alternative approach is to process both SISO decoding in parallel but it requires additional computations for synchronization and/or impacts on error-correction performance [24].…”
Section: Parallelism Analysismentioning
confidence: 99%
“…In [9]- [18], turbo decoders were implemented on GPU targets to benefit from their computing power in order to comply with the LTE required throughputs. This was made possible by exploiting the parallelism within the turbo decoding process (intra-frame parallelism).…”
Section: Introductionmentioning
confidence: 99%
“…In [2,3,4], GPUs are used to realize turbo decoding algorithm for their computing power and programmability. By fully utilizing the enormous parallelism of GPUs, many codeword blocks can be decoded simultaneously.…”
Section: Introductionmentioning
confidence: 99%