2022
DOI: 10.48550/arxiv.2201.00057
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Optimal Representations for Covariate Shift

Abstract: Machine learning systems often experience a distribution shift between training and testing. In this paper, we introduce a simple variational objective whose optima are exactly the set of all representations on which risk minimizers are guaranteed to be robust to any distribution shift that preserves the Bayes predictor, e.g., covariate shifts. Our objective has two components. First, a representation must remain discriminative for the task, i.e., some predictor must be able to simultaneously minimize the sour… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 27 publications
0
4
0
Order By: Relevance
“…Spurious correlations and representation learning. Prior work has often associated poor robustness to spurious correlations with the quality of representations learned by the model [4,6,84] and suggested that the entire model needs to be carefully trained to avoid relying on spurious features [e.g. 85,43,58,107,92,76,56].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Spurious correlations and representation learning. Prior work has often associated poor robustness to spurious correlations with the quality of representations learned by the model [4,6,84] and suggested that the entire model needs to be carefully trained to avoid relying on spurious features [e.g. 85,43,58,107,92,76,56].…”
Section: Discussionmentioning
confidence: 99%
“…Other papers proposed methods based on meta-learning the weights for a weighted loss [80] and group-agnostic adaptive regularization [15,14]. Related methods have been developed in several areas of machine learning, such as ML Fairness [22,30,47,77,2,46], domain adaptation [23,24] and domain generalization [9,65,57,29,84] including works on Invariant Risk Minimization and causality [75,4,53,5].…”
Section: Related Workmentioning
confidence: 99%
“…In addition, both M min cpt and M max cpt have close connections with DG performance as shown in Theorem 4.2 and Theorem 4.3 in Section 4.2. The covariate shift is widely discussed in recent literature [Duchi et al, 2020;Ruan et al, 2021; yet none of them give the quantification with function discrepancy, which favors the analysis of DG performance and shows remarkable properties when H is large (such as the function space supported by current deep models). The concept shift can be considered as the discrepancy between the labeling rule f tr on the training data and the labeling rule f te on the test data.…”
Section: Metrics For Covariate Shift and Concept Shiftmentioning
confidence: 99%
“…Domain adaptation is another closely related problem setting. Domain adaptation (DA) methods require access to labeled source and unlabeled target domains during training and aim to improve target performance via a combination of distribution matching [18,50,43], self-training [46,31], data augmentation [8,38], and other regularizers. DA methods are typically challenging to train and require retraining for every new target domain.…”
Section: Related Workmentioning
confidence: 99%