2013
DOI: 10.1007/978-3-642-40991-2_18
|View full text |Cite
|
Sign up to set email alerts
|

An Analysis of Tensor Models for Learning on Structured Data

Abstract: Abstract. While tensor factorizations have become increasingly popular for learning on various forms of structured data, only very few theoretical results exist on the generalization abilities of these methods. Here, we discuss the tensor product as a principled way to represent structured data in vector spaces for machine learning tasks. By extending known bounds for matrix factorizations, we are able to derive generalization error bounds for the tensor case. Furthermore, we analyze analytically and experimen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 12 publications
(13 citation statements)
references
References 22 publications
0
13
0
Order By: Relevance
“…The reason for feature construction is that the training data determines the highest accuracy that can be achieved. Appropriate feature construction can increase the useful information in the data, which helps to improve the classification accuracy of the model [44][45][46]. Therefore, based on the original features, the accuracy of the classification algorithm is improved by constructing new features.…”
Section: Feature Construction Methodsmentioning
confidence: 99%
“…The reason for feature construction is that the training data determines the highest accuracy that can be achieved. Appropriate feature construction can increase the useful information in the data, which helps to improve the classification accuracy of the model [44][45][46]. Therefore, based on the original features, the accuracy of the classification algorithm is improved by constructing new features.…”
Section: Feature Construction Methodsmentioning
confidence: 99%
“…This allows us to define, in a conceptually simple way, the hypothesis class H G corresponding to the family of linear models whose weights are represented using an arbitrary TN structure G. We then proceed to deriving upper bounds on the VC/pseudodimension and generalization error of the class H G . These bounds follow from a classical result from Warren [64] which was previously used to obtain generalization bounds for neural networks [3], matrix completion [52] and tensor completion [39]. The bounds we derive naturally relate the capacity of H G to the underlying graph structure G through the number of nodes and effective number of parameters of the TN.…”
Section: Introductionmentioning
confidence: 90%
“…Related work Machine learning models using low-rank parametrization of the weights have been investigated (mainly from a practical perspective) for various decomposition models, including low-rank matrices [34,47,65], CP [1,35,6], Tucker [33,15,25,48], tensor train [46,9,42,54,17,51,10,63,66] and PEPS [11]. From a more theoretical perspective, generalization bounds for matrix and tensor completion have been derived in [52,39] (based on the Tucker format for the tensor case). A bound on the VC-dimension of low-rank matrix classifiers was derived in [65] and a bound on the pseudo-dimension of regression functions whose weights have low Tucker rank was given in [48] (for both these cases, we show that our results improve over these previous bounds, see Section 4.2).…”
Section: Introductionmentioning
confidence: 99%
“…Thus tensor factorizations can easily integrate multiple data modalities, reduce dimensionality and identify latent groups in each mode for meaningful summarization of both features and instances as propounded by [36] in medical data analysis. According to Nickel et al [37] tensor as a factorization tool has a collective entity for filtering hidden factors related to massive data. In Reference [38], it is proven that, tensor-based methods is ideal for mitigating personalized tagging and link prediction recommendation.…”
Section: General Frameworkmentioning
confidence: 99%
“…In the Google's wide and deep model [44], a generalized linear model was used to capture latent features in the wide perspective. In this paper, tensor factorization model which is non-linear in nature is chosen to be a platform model as a result of the fact that it has appealing property to efficiently impose structure on the vector space representation of the data as propounded by [37]. We will regard an array of numbers with more than 2 dimensions as a tensor.…”
Section: Multi-task Tensor Factorizationmentioning
confidence: 99%