Proceedings of the 24th International Conference on Machine Learning 2007
DOI: 10.1145/1273496.1273499
|View full text |Cite
|
Sign up to set email alerts
|

Uncovering shared structures in multiclass classification

Abstract: This paper suggests a method for multiclass learning with many classes by simultaneously learning shared characteristics common to the classes, and predictors for the classes in terms of these characteristics. We cast this as a convex optimization problem, using trace-norm regularization and study gradient-based optimization both for the linear case and the kernelized setting.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
267
0

Year Published

2008
2008
2022
2022

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 229 publications
(269 citation statements)
references
References 12 publications
1
267
0
Order By: Relevance
“…Our word prediction formulation of the loss is different from [21] (a pure collaborative filtering model) and [1] (a multi-class classifier), even though our tracenorm regularization term is similar to theirs. Our formulation is, to the best of our knowledge, the first application of the tracenorm regularization to a problem of these characteristics.…”
Section: A Max-margin Factorization Modelmentioning
confidence: 99%
See 3 more Smart Citations
“…Our word prediction formulation of the loss is different from [21] (a pure collaborative filtering model) and [1] (a multi-class classifier), even though our tracenorm regularization term is similar to theirs. Our formulation is, to the best of our knowledge, the first application of the tracenorm regularization to a problem of these characteristics.…”
Section: A Max-margin Factorization Modelmentioning
confidence: 99%
“…Our formulation is, to the best of our knowledge, the first application of the tracenorm regularization to a problem of these characteristics. From [1] we took the optimization framework, although we are using different losses and approximations and we are using BFGS to perform the minimization. Finally, we introduce a unsupervised model on top of the internal representation this formulation produces to discover scenes.…”
Section: A Max-margin Factorization Modelmentioning
confidence: 99%
See 2 more Smart Citations
“…where [15,24], we use the trace norm regularization for simultaneous feature extraction and model learning. As the prediction are dependent on F E H and F G G, we can add the Frobenius norm regularizer in the above EM algorithm to control the magnitude of F E , Fg, H and G. Adding the regularization terms will not change the posterior distribution of the latent variable z.…”
Section: Mixture Of Experts Model With the Trace Norm Regularizationmentioning
confidence: 99%