2019
DOI: 10.48550/arxiv.1906.01614
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Confidence Regions in Wasserstein Distributionally Robust Estimation

Abstract: Wasserstein distributionally robust optimization (DRO) estimators are obtained as solutions of min-max problems in which the statistician selects a parameter minimizing the worst-case loss among all probability models within a certain distance (in a Wasserstein sense) from the underlying empirical measure. While motivated by the need to identify model parameters (or) decision choices that are robust to model uncertainties and misspecification, the Wasserstein DRO estimators recover a wide range of regularized … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 31 publications
0
6
0
Order By: Relevance
“…, where the second inequality follows from Assumption 5 and [23, Lemma 2] (see also [24]). The desired result then readily follows from inequalities (15) and (17).…”
Section: A2 Proof Of Theoremmentioning
confidence: 84%
See 1 more Smart Citation
“…, where the second inequality follows from Assumption 5 and [23, Lemma 2] (see also [24]). The desired result then readily follows from inequalities (15) and (17).…”
Section: A2 Proof Of Theoremmentioning
confidence: 84%
“…) results in estimators with a finite sample certificate-type guarantee (cf. [15]). This larger choice of the radius ζ n also yields estimators with the conventional O p (n −1/2 ) rate of convergence when we use parametric regression methods to estimate the function f * .…”
Section: Wasserstein-based Ambiguity Setsmentioning
confidence: 99%
“…where C 1 (θ) is the standard deviation of the loss function, Var 0 (l(θ, ξ)), multiplied by a constant that depends on φ. Similarly, if D is the Wasserstein distance and η is of order 1/n, (4) holds with C 1 (θ) being the gradient norm or the Lipschitz norm (Blanchet et al 2019a, Gao et al 2017, Blanchet et al 2019b). Furthermore, Staib and Jegelka (2019), which is perhaps closest to our work, studies MMD as the DRO statistical distance and derives a high-probability bound for Z(θ) similar to the RHS of (4), with C 1 (θ) being the reproducing kernel Hilbert space (RKHS) norm of l(θ, •).…”
Section: Variability Regularizationmentioning
confidence: 99%
“…As an effective way for the decision-making under uncertainty without fully knowing the probability distribution, distributionally robust optimization has attracted much attention [8,10,12,16,21,22,41,43,46]. Interested readers are referred to [34] for a comprehensive review.…”
Section: Relevant Literaturementioning
confidence: 99%
“…(a) If we focus on convergence of the optimal objective value instead of probability distribution, then the Wasserstein radius should be chosen in the order of O(N − 1 2 ) rather than O(N − 1 n ) shown in [14,15]. (b) The asymptotic rate O(N − 1 2 ) has been observed in [8] from a different angle. However, our rate is non-asymptotic.…”
Section: Proofmentioning
confidence: 99%