2022
DOI: 10.3389/frai.2022.900943
|View full text |Cite
|
Sign up to set email alerts
|

Judgment aggregation, discursive dilemma and reflective equilibrium: Neural language models as self-improving doxastic agents

Abstract: Neural language models (NLMs) are susceptible to producing inconsistent output. This paper proposes a new diagnosis as well as a novel remedy for NLMs' incoherence. We train NLMs on synthetic text corpora that are created by simulating text production in a society. For diagnostic purposes, we explicitly model the individual belief systems of artificial agents (authors) who produce corpus texts. NLMs, trained on those texts, can be shown to aggregate the judgments of individual authors during pre-training accor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
4
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(7 citation statements)
references
References 76 publications
3
4
0
Order By: Relevance
“…Self-training improves logical alignment of credences. It brings down, consistent with previous findings [ 61 ], the frequency of transitivity violations ( Fig 3 ) and improves further consistency metrics (see also S3 Fig ).…”
Section: Resultssupporting
confidence: 92%
See 4 more Smart Citations
“…Self-training improves logical alignment of credences. It brings down, consistent with previous findings [ 61 ], the frequency of transitivity violations ( Fig 3 ) and improves further consistency metrics (see also S3 Fig ).…”
Section: Resultssupporting
confidence: 92%
“…And in case self-training doesn’t improve a specific metric (such as logical alignment with entailment relation, reflexivity, or complementarity), we can explain the lack of improvement, namely with reference to low initial levels, or by means of data segmentation and in-depth analysis. The observed logical benefits of self-training are consistent with [ 61 , 73 ].…”
Section: Discussionsupporting
confidence: 89%
See 3 more Smart Citations