2019
DOI: 10.48550/arxiv.1911.09105
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On Universal Features for High-Dimensional Learning and Inference

Shao-Lun Huang,
Anuran Makur,
Gregory W. Wornell
et al.

Abstract: We consider the problem of identifying universal low-dimensional features from high-dimensional data for inference tasks in settings involving learning. For such problems, we introduce natural notions of universality and we show a local equivalence among them. Our analysis is naturally expressed via information geometry, and represents a conceptually and computationally useful analysis. The development reveals the complementary roles of the singular value decomposition, Hirschfeld-Gebelein-Rényi maximal correl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
21
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(22 citation statements)
references
References 89 publications
(206 reference statements)
1
21
0
Order By: Relevance
“…Here, H-score is a metric determining whether certain features are informative for the task (more formally introduced by Bao et al 2019) 1 . Its derivation is rooted in the maximal correlation interpretation of deep neural networks (Huang et al 2019a), where given the extracted features, the formula of H-score is actually the optimal network performance under an information theoretic measurement. Previous work also extended H-score to a transferability metric and verified its effectiveness with extensive experiments (Bao et al 2019;Ibrahim, Ponomareva, and Mazumder 2022), suggesting the potential of H-score in transfer learning.…”
Section: Related Work Maximal Correlation Regression and H-scorementioning
confidence: 99%
“…Here, H-score is a metric determining whether certain features are informative for the task (more formally introduced by Bao et al 2019) 1 . Its derivation is rooted in the maximal correlation interpretation of deep neural networks (Huang et al 2019a), where given the extracted features, the formula of H-score is actually the optimal network performance under an information theoretic measurement. Previous work also extended H-score to a transferability metric and verified its effectiveness with extensive experiments (Bao et al 2019;Ibrahim, Ponomareva, and Mazumder 2022), suggesting the potential of H-score in transfer learning.…”
Section: Related Work Maximal Correlation Regression and H-scorementioning
confidence: 99%
“…− p(x)p(y), for discrete random variables x and y. It is shown in [27] that, under the assumption of sufficiently small B, the solution of the cross-entropy loss coincides with the following solution of the matrix decomposition:…”
Section: H-scorementioning
confidence: 99%
“…However, the key drawback of the H-score is that the optimality is based on the least squares objective, which is rarely used for classification. As proven in [27], the least squares solution is a valid approximation to the cross-entropy classification loss only when label y and input x are weakly dependent, which is clearly not the case in general.…”
Section: H-scorementioning
confidence: 99%
See 1 more Smart Citation
“…The advantage of using bivariate distributions is that we can have probabilistic insights on network analysis. A very interesting recent work on the embedding problem for a bipartite network [35] also used a bivariate distribution to characterize a useritem network. There they showed that the optimal embedding vectors can be interpreted as conditional expectations.…”
Section: Equivalence Of the Embedding Problem In Sampled Graphsmentioning
confidence: 99%