2015 IEEE Global Conference on Signal and Information Processing (GlobalSIP) 2015
DOI: 10.1109/globalsip.2015.7418298
|View full text |Cite
|
Sign up to set email alerts
|

Mobile GPU accelerated digital predistortion on a software-defined mobile transmitter

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
4
1
1

Relationship

3
3

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 16 publications
0
5
0
Order By: Relevance
“…Moreover, the LS problem is often poorly conditioned [4]. In [10], a mobile graphics processing units (GPU) was used to implement the polynomial DPD with I/Q imbalance correction from [4]. This GPU implementation used floating-point and was able to avoid the challenges associated with the dynamic range requirements for memory polynomials.…”
Section: Introductionmentioning
confidence: 99%
“…Moreover, the LS problem is often poorly conditioned [4]. In [10], a mobile graphics processing units (GPU) was used to implement the polynomial DPD with I/Q imbalance correction from [4]. This GPU implementation used floating-point and was able to avoid the challenges associated with the dynamic range requirements for memory polynomials.…”
Section: Introductionmentioning
confidence: 99%
“…One of the examples of such a problem is the delay time due to finite speed of processor calculations and its relation to maximal network latency (which is defined as time it takes data to travel from one point to another) of less than 1 ms in 5G, which we mentioned in the introduction. In [93][94][95], the delay of DPD, depending on the implementation (FPGA, GPU, ASIC), is evaluated to be tens to hundreds of microseconds. Taking into account the fact that signal may have to pass through multiple DPD stages and through the whole data transfer system, it is possible DPD delay will be a major bottleneck in overall system latency.…”
Section: Discussionmentioning
confidence: 99%
“…However, few DPD implementations have been done on such general purpose processors, especially mobile processors considering our DPD design targets mobile transmitters. Our previous work [19], to the best of the authors' knowledge, is the first CUDA-based DPD implementation on GPU, and this paper extends from our previous mobile GPU based DPD implementation with further design optimization and thus higher data rate, and details another embedded CPU-based design for comparison. [20] also proposes an alternative implementation on mobile GPU using OpenCL.…”
Section: Introductionmentioning
confidence: 91%
“…On desktop GPUs which support CUDA, we can invoke even more threads and thread-blocks to realize data parallelism on thousands of CUDA cores, and can perform multi-stream scheduling for pipelining CPU-GPU memory copy and kernel execution as an alternative technique of zero copy to resolve memory copy overhead. Our previous work [19] has discussed DPD performance on desktop GPUs as reference. On desktop CPUs, we can take advantage of even longer SIMD instructions, such as SSE and AVX, with 256-bit or 512-bit registers, and generate more OpenMP threads on more CPU cores for higher performance.…”
Section: Design Summary and Comparisonmentioning
confidence: 99%
See 1 more Smart Citation