Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1214
|View full text |Cite
|
Sign up to set email alerts
|

Gromov-Wasserstein Alignment of Word Embedding Spaces

Abstract: Cross-lingual or cross-domain correspondences play key roles in tasks ranging from machine translation to transfer learning. Recently, purely unsupervised methods operating on monolingual embeddings have become effective alignment tools. Current state-of-theart methods, however, involve multiple steps, including heuristic post-hoc refinement strategies. In this paper, we cast the correspondence problem directly as an optimal transport (OT) problem, building on the idea that word embeddings arise from metric re… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
200
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 220 publications
(201 citation statements)
references
References 29 publications
1
200
0
Order By: Relevance
“…Furthermore, there is a surprising lack of testing BLI results for statistical significance. The mismatches in evaluation yield partial conclusions and inconsistencies: on the one hand, some unsupervised models (Artetxe et al, 2018b;Hoshen and Wolf, 2018;Alvarez-Melis and Jaakkola, 2018) are reported to outperform (or at least perform on par with) previous best-performing supervised CLE models (Artetxe et al, 2017;Smith et al, 2017). On the other hand, the most recent supervised approaches (Doval et al, 2018; report further performance gains, surpassing unsupervised models.…”
Section: Introductionmentioning
confidence: 99%
“…Furthermore, there is a surprising lack of testing BLI results for statistical significance. The mismatches in evaluation yield partial conclusions and inconsistencies: on the one hand, some unsupervised models (Artetxe et al, 2018b;Hoshen and Wolf, 2018;Alvarez-Melis and Jaakkola, 2018) are reported to outperform (or at least perform on par with) previous best-performing supervised CLE models (Artetxe et al, 2017;Smith et al, 2017). On the other hand, the most recent supervised approaches (Doval et al, 2018; report further performance gains, surpassing unsupervised models.…”
Section: Introductionmentioning
confidence: 99%
“…Still, one should keep in mind that it has also been shown that there can be stability problems with SHAP. 417 …”
Section: How To Interpret the Results: Avoiding The Clever Hansmentioning
confidence: 99%
“…Thanks to libraries like tf-explain 420 and keras-vis, 421 appealing visualizations of model explanations are often only one function call away, but one should be aware that there are many caveats wherefore some sanity checks (such as randomization tests or addition of noise) should be used before relying on such a model interpretation. 417 , 422…”
Section: How To Interpret the Results: Avoiding The Clever Hansmentioning
confidence: 99%
“…If the weights of alignment are not important, the neural network without attention mechanism may also effectively detect parallel sentences since all alignments have the same contribution. However, the alignment deeply depends on linguistics and context [ 23 25 ]. For example, the English word “bearing” means multiple Chinese words such as “chengzhou,” “baochi,” and “zhoucheng” in a different context.…”
Section: Resultsmentioning
confidence: 99%