2021
DOI: 10.1007/978-3-030-82099-2_25
|View full text |Cite
|
Sign up to set email alerts
|

Finding Fuzziness in Neural Network Models of Language Processing

Abstract: Humans often communicate by using imprecise language, suggesting that fuzzy concepts with unclear boundaries are prevalent in language use. In this paper, we test the extent to which models trained to capture the distributional statistics of language show correspondence to fuzzy-membership patterns. Using the task of natural language inference, we test a recent state of the art model on the classical case of temperature, by examining its mapping of temperature data to fuzzyperceptions such as cool, hot, etc. W… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
9
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(9 citation statements)
references
References 13 publications
0
9
0
Order By: Relevance
“…Most of the NLI models also captured this effect, likely because they are able to encode typicality relations and use these relations for generalization (Han et al, 2022; Misra et al, 2022). However, GPT-DaVinci and BART-MNLI failed to do so.…”
Section: Empirical Regularitiesmentioning
confidence: 99%
See 1 more Smart Citation
“…Most of the NLI models also captured this effect, likely because they are able to encode typicality relations and use these relations for generalization (Han et al, 2022; Misra et al, 2022). However, GPT-DaVinci and BART-MNLI failed to do so.…”
Section: Empirical Regularitiesmentioning
confidence: 99%
“…They also fail at replicating many of the other effects documented in the psychology and cognitive science literatures. This indicates that even though LLMs may possess the type of knowledge necessary for inductive reasoning, for example, knowledge of category membership relations (Misra et al, 2022), they do not possess the reasoning algorithms necessary to generate human-like behavior.…”
mentioning
confidence: 99%
“…If the model achieves a good performance, then one can infer that the LM representation encodes the target linguistic knowledge (Tenney et al, 2019a,b;Hewitt and Liang, 2019;Wu et al, 2020;Vulić et al, 2020;Sorodoc et al, 2020;Ettinger, 2020;Geiger et al, 2021;Koto et al, 2021;Chersoni et al, 2021a;Conia and Navigli, 2022;Kim and Linzen, 2020;Arps et al, 2022;Misra et al, 2022).…”
Section: Probing Linguistic Knowledge In Language Modelsmentioning
confidence: 99%
“…Large language models (LLMs) have attained an extraordinary ability to harness natural language for solving diverse problems (Devlin et al, 2018), often without the need for finetuning Sanh et al, 2021). Moreover, LLMs have demonstrated the capacity to excel at realworld problems, such as mathematics (Lewkowycz et al, 2022), scientific question answering (Sadat and Caragea, 2022), predicting brain responses (Schrimpf et al, 2021), and classifying proteins and chemical compounds (Taylor et al, 2022).…”
Section: Introductionmentioning
confidence: 99%