Proceedings of the Thirtieth Annual ACM-SIAM Symposium on Discrete Algorithms 2019
DOI: 10.1137/1.9781611975482.172
|View full text |Cite
|
Sign up to set email alerts
|

Relative Error Tensor Low Rank Approximation

Abstract: Large language models have become ubiquitous in modern life, finding applications in various domains such as natural language processing, language translation, and speech recognition. Recently, a breakthrough work [Zhao, Panigrahi, Ge, and Arora Arxiv 2023] explains the attention model from probabilistic context-free grammar (PCFG). One of the central computation task for computing probability in PCFG is formulating a particular tensor low rank approximation problem, we can call it tensor cycle rank. Given an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
26
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 41 publications
(26 citation statements)
references
References 161 publications
0
26
0
Order By: Relevance
“…Other sampling techniques for the HOSVD and the HOOI algorithms can be found in [33], [34], [37]. The proposed algorithms in [33] are based on the sparsification idea where in the first step, a sparse tensor is generated from the original data tensor.…”
Section: Algorithm 11: Randomized Higher Order Interpolatory Decomposmentioning
confidence: 99%
See 2 more Smart Citations
“…Other sampling techniques for the HOSVD and the HOOI algorithms can be found in [33], [34], [37]. The proposed algorithms in [33] are based on the sparsification idea where in the first step, a sparse tensor is generated from the original data tensor.…”
Section: Algorithm 11: Randomized Higher Order Interpolatory Decomposmentioning
confidence: 99%
“…The idea of the stochastic gradient descent is used in [37], where at each iteration only a subtensor of the original tensor is considered. The randomized algorithms proposed in [34] are based on sampling fibers instead of components.…”
Section: Algorithm 11: Randomized Higher Order Interpolatory Decomposmentioning
confidence: 99%
See 1 more Smart Citation
“…Aggour et al [1] introduce adaptive sketching for CP decomposition. Song et al [51] discuss the theoretical relative error of various tensor decompositions based on sketching. The work by Cheng et al [13] and Larsen and Kolda [31] accelerate CP-ALS based on leverage score sampling.…”
Section: Previous Workmentioning
confidence: 99%
“…We build our even faster algorithm via sketching techniques. Sketching has been successfully applied to speed up different problems, such as linear programs [LSZ19], clustering [CEM + 15, SYZ18], low rank approximations [CW13, NN13, BW14, CW15a, RSW16, SWZ17], linear regression [CW13, NN13, CLM + 15, CW15b, PSW17, ALS + 18], total least regression [DSWY19], tensor regression [LHW17, DSSW18] and tensor decomposition [WTSA15, SWZ16,SWZ19]. Readers may refer to [Woo14] for a comprehensive survey on sketching technique.…”
Section: Related Workmentioning
confidence: 99%