2022
DOI: 10.1007/s10586-022-03805-x
|View full text |Cite
|
Sign up to set email alerts
|

Improving Oversubscribed GPU Memory Performance in the PyTorch Framework

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 19 publications
0
2
0
Order By: Relevance
“…The dataset comprised the training (800 images) and test (200 images) sets. The former was used to train the model using the Pytorch framework [21] on an NVIDIA GeForce GTX 3090Ti GPU for tting the parameters of the model. The latter was used to evaluate the performance of the model by comparing its results with manually measured results.…”
Section: Model Trainingmentioning
confidence: 99%
“…The dataset comprised the training (800 images) and test (200 images) sets. The former was used to train the model using the Pytorch framework [21] on an NVIDIA GeForce GTX 3090Ti GPU for tting the parameters of the model. The latter was used to evaluate the performance of the model by comparing its results with manually measured results.…”
Section: Model Trainingmentioning
confidence: 99%
“…Additionally, UM supports GPU memory oversubscription, i.e., GPU kernels access more data than the GPU memory can hold, significantly enhancing programming portability and productivity for memory-demanding workloads. UM technologies have been adopted by HPC frameworks such as Raja [6], Kokkos [9], and Trilinos [16] for writing portable applications on today's and future's major HPC platforms, and by deep learning frameworks [12,22,34]. However, even with active research and improvement by vendors and research community [3,18,23,42], current UM technologies cause significant, or even prohibitive, performance degradation [25,26,46].…”
Section: Introductionmentioning
confidence: 99%