2017
DOI: 10.48550/arxiv.1705.08056
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Ambiguity set and learning via Bregman and Wasserstein

Abstract: Construction of ambiguity set in robust optimization relies on the choice of divergences between probability distributions. In distribution learning, choosing appropriate probability distributions based on observed data is critical for approximating the true distribution. To improve the performance of machine learning models, there has recently been interest in designing objective functions based on Lp-Wasserstein distance rather than the classical Kullback-Leibler (KL) divergence. In this paper, we derive con… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 25 publications
0
5
0
Order By: Relevance
“…Remark 3 (Comparisons with Wasserstein-Bregman divergence). We notice that formulation (i) is similar but different with divergence functionals defined in [13,33]. In our formulation, joint density π is the optimal one from the L 2 -Wasserstein space; while in [13] or [33], joint density π solves the related linear programming problem w.r.t.…”
Section: 3mentioning
confidence: 99%
See 3 more Smart Citations
“…Remark 3 (Comparisons with Wasserstein-Bregman divergence). We notice that formulation (i) is similar but different with divergence functionals defined in [13,33]. In our formulation, joint density π is the optimal one from the L 2 -Wasserstein space; while in [13] or [33], joint density π solves the related linear programming problem w.r.t.…”
Section: 3mentioning
confidence: 99%
“…There are joint works in the literature between optimal transport and information geometry to study Bregman divergences [2,7,13,22,23,25,33,34]. In particular, [13,33,34] apply linear programming formulations of optimal transport.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…Note that for two continuous distributions P 1 and P 2 having probability density functions p 1 (x) and p 2 (x), respectively, the Bregman divergence can be defined as (Guo et al, 2017;Jones & Byrne, 1990)…”
Section: Proof Of Optimal Rewards On Continuous Controlsmentioning
confidence: 99%