Proceedings of the ACM International Conference on Supercomputing 2021
DOI: 10.1145/3447818.3460692
|View full text |Cite
|
Sign up to set email alerts
|

An optimized tensor completion library for multiple GPUs

Abstract: Tensor computations are gaining wide adoption in big data analysis and artificial intelligence. Among them, tensor completion is used to predict the missing or unobserved value in tensors. The decomposition-based tensor completion algorithms have attracted significant research attention since they exhibit better parallelization and scalability. However, existing optimization techniques for tensor completion cannot sustain the increasing demand for applying tensor completion on ever larger tensor data. To addre… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(12 citation statements)
references
References 37 publications
0
12
0
Order By: Relevance
“…Optimizing sparse MTTKRP has been the subject of several prior studies, which propose sparse tensor formats along with parallel algorithms to process/analyze the data. List-based formats, such as F-COO [30], GenTen [39], and TB-COO [12], explicitly store the multi-dimensional coordinates of each non-zero element. To reduce atomic operations, these formats store multiple mode-specific copies of the tensor and/or extra scheduling information, which substantially increases their memory footprint.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Optimizing sparse MTTKRP has been the subject of several prior studies, which propose sparse tensor formats along with parallel algorithms to process/analyze the data. List-based formats, such as F-COO [30], GenTen [39], and TB-COO [12], explicitly store the multi-dimensional coordinates of each non-zero element. To reduce atomic operations, these formats store multiple mode-specific copies of the tensor and/or extra scheduling information, which substantially increases their memory footprint.…”
Section: Related Workmentioning
confidence: 99%
“…Segmented scan and reduction [42,52] have been used to reduce the synchronization cost of sparse workloads [6,12,29,51,53]. Prior studies apply these primitives to mode-specific formats with delineated and/or sorted groups of non-zero elements according to the target mode.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations