Recent large language models (LLMs) and LLM-driven chatbots, such as ChatGPT, have sparked debate regarding whether these artificial systems can develop human-like linguistic capacities. We examined this issue by investigating whether ChatGPT resembles humans in its ability to enrich literal meanings of utterances with pragmatic implicatures. Humans not only distinguish implicatures from truth-conditional meanings of utterances but also compute implicatures contingent on the communicative context. In three preregistered experiments (https://osf.io/4bcx9/), we assessed whether ChatGPT resembles humans in the computation of pragmatic implicatures. Experiment 1 investigated generalized conversational implicatures (GCIs); for example, the utterance “She walked into the bathroom. The window was open.” has the conversational implicature that the window is located in the bathroom, while the truth-conditional (literal) meaning of the utterance allows for the possibility that the window is located elsewhere. Humans demonstrate their ability to distinguish GCIs from the truth-conditional meanings by inhibiting the computation of GCIs when explicitly instructed to focus on the literal sense of the utterances. We tested whether ChatGPT could also inhibit the computation of GCIs as humans do. Experiment 2 and Experiment 3 investigated whether the communicative context modulates how ChatGPT computes a specific type of GCIs, namely scalar implicatures (SIs). For humans, the sentence “Julie had found a crab or a starfish” implies that Julie did not have both a crab and a starfish, even though the sentence’s literal meaning allows for this possibility. Moreover, this implicature is argued to be more available when the word “or” is in the information focus, e.g. as a reply to the question “What had Julie found?” than in the information background, e.g. as a reply to the question “Who had found a crab or a starfish?”. Experiment 2 tested whether ChatGPT shows similar sensitivity to information structure when computing SIs. Experiment 3 focused on a different contextual aspect, investigating whether face-threatening and face-boosting contexts have different effects on how ChatGPT computes SIs. Previous research has shown that humans interlocutors compute more SIs in face-boosting contexts, e.g. interpreting the utterance “Some people loved your poem.” as saying “Not all people loved your poem.” but not so much when they are in face-threatening contexts; and we tested whether ChatGPT exhibits a similar tendency. In all three experiments, ChatGPT did not display human-like flexibility in switching between pragmatic and semantic processing and failed to show the well-established effects of communicative context on the SI rate. Overall, our experiments demonstrate that although ChatGPT parallels or even surpasses humans in many linguistic tasks, it still does not closely resemble human beings in the computation of GCIs.
Though lying is a common deceptive behavior that is acquired very early in life, the exact definition of it varies across researchers. The debates pertaining to the definition of lying suggest individual differences in the moral evaluation of deceptive verbal behaviors. It is possible that individuals with different personality traits conceptually distinguish lies from other types of deception in a systematic pattern, however, this topic remains largely unexplored. This article suggests possible links between personality traits and the ideology of what constitutes a lie, linking linguistic inquires to personality psychology.
This paper presents a series of three experiments investigating the processing of nested epistemic expressions, utterances containing two epistemic modals in one clause, such as “he certainly may have forgotten”. While some linguists claim that in a nested epistemic expression, one modal is semantically embedded within the scope of the other modal based on the word order, it is possible that in daily conversation, the scope of nested modals may not be thoroughly processed, leading to a “good-enough” interpretation that is not sensitive to the word order of the two modals. This study used probability judgment tests to investigate people’s interpretation of nested epistemic expressions, and the effect of word order was not observed. This result fails to support the scope account of the nested epistemic expressions and suggests a holistic processing mechanism in line with the “good-enough” processing framework.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.