Findings of the Association for Computational Linguistics: NAACL 2022 2022
DOI: 10.18653/v1/2022.findings-naacl.156
|View full text |Cite
|
Sign up to set email alerts
|

Beyond Distributional Hypothesis: Let Language Models Learn Meaning-Text Correspondence

Abstract: The logical negation property (LNP), which implies generating different predictions for semantically opposite inputs (p is true iff ¬p is false), is an important property that a trustworthy language model must satisfy. However, much recent evidence shows that large-size pre-trained language models (PLMs) do not satisfy this property. In this paper, we perform experiments using probing tasks to assess PLMs' LNP understanding. Unlike previous studies that only examined negation expressions, we expand the boundar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
23
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(23 citation statements)
references
References 27 publications
0
23
0
Order By: Relevance
“…Similar to other LLMs, ChatGPT also suffers from a lack of consistency [32,33]. The model frequently alters its decisions when presented with a paraphrased sentence, revealing self-contradictory behavior.…”
Section: Related Workmentioning
confidence: 99%
“…Similar to other LLMs, ChatGPT also suffers from a lack of consistency [32,33]. The model frequently alters its decisions when presented with a paraphrased sentence, revealing self-contradictory behavior.…”
Section: Related Workmentioning
confidence: 99%
“…To alleviate the issue, several works adopted data augmentation to train a model with abundant data containing negation expressions (Asai and Hajishirzi, 2020;Hosseini et al, 2021). Jang et al (2022b) introduced the meaning-matching task to enhance PLMs' textual understanding ability and observed performance improvements.…”
Section: Related Workmentioning
confidence: 99%
“…A correct behaviour is a crucial aspect in deciding models' trustworthiness by improving the certification 1 process (Jang et al, 2022a). In this regard, we mainly investigate the trustworthiness of ChatGPT in terms of logically consistent behaviour.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In previous work, we showed how natural language notions of context can be modelled by the mathematics of quantum contextuality [36,37,34]. In particular, we modelled anaphoric context in [28].…”
Section: Introductionmentioning
confidence: 99%