2023
DOI: 10.1371/journal.pone.0281372
|View full text |Cite
|
Sign up to set email alerts
|

Probabilistic coherence, logical consistency, and Bayesian learning: Neural language models as epistemic agents

Abstract: It is argued that suitably trained neural language models exhibit key properties of epistemic agency: they hold probabilistically coherent and logically consistent degrees of belief, which they can rationally revise in the face of novel evidence. To this purpose, we conduct computational experiments with rankers: T5 models [Raffel et al. 2020] that are pretrained on carefully designed synthetic corpora. Moreover, we introduce a procedure for eliciting a model’s degrees of belief, and define numerical metrics t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 58 publications
(59 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?