2021
DOI: 10.48550/arxiv.2110.04020
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pathologies in priors and inference for Bayesian transformers

Abstract: In recent years, the transformer has established itself as a workhorse in many applications ranging from natural language processing to reinforcement learning. Similarly, Bayesian deep learning has become the gold-standard for uncertainty estimation in safety-critical applications, where robustness and calibration are crucial. Surprisingly, no successful attempts to improve transformer models in terms of predictive uncertainty using Bayesian inference exist. In this work, we study this curiously underpopulated… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 5 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?