2024
DOI: 10.1038/s41598-024-79531-8
|View full text |Cite
|
Sign up to set email alerts
|

Testing AI on language comprehension tasks reveals insensitivity to underlying meaning

Vittoria Dentella,
Fritz Günther,
Elliot Murphy
et al.

Abstract: Large Language Models (LLMs) are recruited in applications that span from clinical assistance and legal support to question answering and education. Their success in specialized tasks has led to the claim that they possess human-like linguistic capabilities related to compositional understanding and reasoning. Yet, reverse-engineering is bound by Moravec's Paradox, according to which easy skills are hard. We systematically assess 7 state-of-the-art models on a novel benchmark. Models answered a series of compr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
references
References 47 publications
0
0
0
Order By: Relevance