ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2021
DOI: 10.1109/icassp39728.2021.9414891
|View full text |Cite
|
Sign up to set email alerts
|

Embedding Semantic Hierarchy in Discrete Optimal Transport for Risk Minimization

Abstract: The widely-used cross-entropy (CE) loss-based deep networks achieved significant progress w.r.t. the classification accuracy. However, the CE loss can essentially ignore the risk of misclassification which is usually measured by the distance between the prediction and label in a semantic hierarchical tree. In this paper, we propose to incorporate the risk-aware inter-class correlation in a discrete optimal transport (DOT) training framework by configuring its ground distance matrix. The ground distance matrix … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
1

Relationship

2
5

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 20 publications
0
4
0
Order By: Relevance
“…Kang et al [14] proposed CDD, which incorporates class labels into MMD by estimating the labels of the target domain through alternating clustering, thus achieving alignment of class-conditional distributions. Ge et al [16] used the Wasserstein distance as a measure of distribution divergence and incorporated risk-aware interclass correlations into the training framework by configuring the distance matrix of the Discrete Optimal Transport (DOT) training framework. This method can better capture the similarity between the source and target domains, but adjusting the size of the distance matrix and the network structure requires some experience and expertise, and the model performance may be limited when the dataset is small.…”
Section: Related Workmentioning
confidence: 99%
“…Kang et al [14] proposed CDD, which incorporates class labels into MMD by estimating the labels of the target domain through alternating clustering, thus achieving alignment of class-conditional distributions. Ge et al [16] used the Wasserstein distance as a measure of distribution divergence and incorporated risk-aware interclass correlations into the training framework by configuring the distance matrix of the Discrete Optimal Transport (DOT) training framework. This method can better capture the similarity between the source and target domains, but adjusting the size of the distance matrix and the network structure requires some experience and expertise, and the model performance may be limited when the dataset is small.…”
Section: Related Workmentioning
confidence: 99%
“…Kang et al [14] proposed CDD, which incorporates class labels into MMD by estimating the labels of the target domain through alternating clustering, thus achieving alignment of class-conditional distributions. Ge et al [16] used the Wasserstein distance as a measure of distribution divergence and incorporated risk-aware inter-class correlations into the training framework by configuring the distance matrix of the Discrete Optimal Transport (DOT) training framework. This method can better capture the similarity between the source and target domains, but adjusting the size of the distance matrix and the network structure requires some experience and expertise, and the model performance may be limited when the dataset is small.…”
Section: Related Workmentioning
confidence: 99%
“…In addition, the Wasserstein distance (Liu et al, 2020d;Han et al, 2020;Liu et al, 2019c;Ge et al, 2021b;Liu et al, 2019c), a.k.a. optimal transportation distance or earth mover's distance (Liu et al, 2020b,c), could be another alternative to measure the distribution divergence.…”
Section: Statistic Divergence Alignmentmentioning
confidence: 99%