2021
DOI: 10.48550/arxiv.2109.14723
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

BeliefBank: Adding Memory to a Pre-Trained Language Model for a Systematic Notion of Belief

Abstract: Although pretrained language models (PTLMs) contain significant amounts of world knowledge, they can still produce inconsistent answers to questions when probed, even after specialized training. As a result, it can be hard to identify what the model actually "believes" about the world, making it susceptible to inconsistent behavior and simple errors. Our goal is to reduce these problems. Our approach is to embed a PTLM in a broader system that also includes an evolving, symbolic memory of beliefs -a BeliefBank… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(6 citation statements)
references
References 16 publications
0
6
0
Order By: Relevance
“…[7] evaluate and improve factual consistency of pre-trained LMs across paraphrasings of factual statements. [15] consider the responses of a pre-trained LM to a stream of questions, and evaluate and improve the consistency and accuracy of its answers over time. [16] collect counterfactual instances to evaluate the overreliance of NLP models on spurious attributes.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…[7] evaluate and improve factual consistency of pre-trained LMs across paraphrasings of factual statements. [15] consider the responses of a pre-trained LM to a stream of questions, and evaluate and improve the consistency and accuracy of its answers over time. [16] collect counterfactual instances to evaluate the overreliance of NLP models on spurious attributes.…”
Section: Related Workmentioning
confidence: 99%
“…In computer vision, cross-task consistency has been of some interest for classical tasks [26], while in natural language processing past work has studied consistency between tasks like question-answering [15]. However, in vision-and-language research, much work has focused on…”
Section: Introductionmentioning
confidence: 99%
“…Selected Tasks GLUECons contains tasks ranging over five different types of problems categorized based on the type of available knowledge. This includes 1) Classification with label dependencies: Mutual exclusivity in multiclass classification using MNIST (LeCun et al 1998) and Hierarchical image classification using CIFAR 100 (Krizhevsky and Hinton 2009), 2) Self-Consistency in decisions: What-If Question Answering (Tandon et al 2019), Natural Language Inference (Bowman et al 2015), BeliefBank (Kassner et al 2021), 3) Consistency with external knowledge: Entity and Relation Extraction using CONLL2003 (Sang and De Meulder 2003), 4) Structural Consistency: BIO Tagging, 5) Constraints in (un/semi)supervised setting: MNIST Arithmetic and Sudoku. These tasks either use existing datasets or are extensions of existing tasks, reformulated so that the usage of knowledge is applicable to them.…”
Section: Knowledge Integration Solutionmentioning
confidence: 99%
“…We use off-the-shelf ILP tools that perform an efficient search and offer a natural way to integrate constraints. However, constraints should be converted to a linear form to be able to exploit these tools Kordjamshidi, Roth, and Wu 2015;Kordjamshidi et al 2016).…”
Section: Constraint Integration In Prior Researchmentioning
confidence: 99%
“…Similar to the efforts done for overcoming lack of consistency under different paraphrases, Hase et al (2021) add another loss term to their objective function to minimize the error across entailed data. On the other hand, Kassner et al (2021) use a feedback mechanism that issues relevant information from a symbolic memory of beliefs as input to the system during test-time in order to improve consistency under entailment.…”
Section: Consistencymentioning
confidence: 99%