Cognitive control refers to adjusting thoughts and actions when confronted with conflict during information processing. We tested whether this ability is causally linked to performance on certain language and memory tasks by using cognitive control training to systematically modulate people's ability to resolve information-conflict across domains. Different groups of subjects trained on 1 of 3 minimally different versions of an n-back task: n-back-with-lures (High-Conflict), n-back-without-lures (Low-Conflict), or 3-back-without-lures (3-Back). Subjects completed a battery of recognition memory and language processing tasks that comprised both high- and low-conflict conditions before and after training. We compared the transfer profiles of (a) the High- versus Low-Conflict groups to test how conflict resolution training contributes to transfer effects, and (b) the 3-Back versus Low-Conflict groups to test for differences not involving cognitive control. High-Conflict training-but not Low-Conflict training-produced discernable benefits on several untrained transfer tasks, but only under selective conditions requiring cognitive control. This suggests that the conflict-focused intervention influenced functioning on ostensibly different outcome measures across memory and language domains. 3-Back training resulted in occasional improvements on the outcome measures, but these were not selective for conditions involving conflict resolution. We conclude that domain-general cognitive control mechanisms are plastic, at least temporarily, and may play a causal role in linguistic and nonlinguistic performance. (PsycINFO Database Record
Few studies have addressed the neural computations underlying decisions made for others despite the importance of this ubiquitous behavior. Using participant-specific behavioral modeling with univariate and multivariate fMRI approaches, we investigated the neural correlates of decision-making for self and other in two independent tasks, including intertemporal and risky choice. Modeling subjective valuation indicated that participants distinguished between themselves and others with dissimilar preferences. Activity in the dorsomedial prefrontal cortex (dmPFC) and ventromedial prefrontal cortex (vmPFC) was consistently modulated by relative subjective value. Multi-voxel pattern analysis indicated that activity in the dmPFC uniquely encoded relative subjective value and generalized across self and other and across both tasks. Furthermore, agent cross-decoding accuracy between self and other in the dmPFC was related to self-reported social attitudes. These findings indicate that the dmPFC emerges as a medial prefrontal node that utilizes a task-invariant mechanism for computing relative subjective value for self and other.
Behavioral evidence and theory suggest gesture and language processing may be part of a shared cognitive system for communication. While much research demonstrates both gesture and language recruit regions along perisylvian cortex, relatively less work has tested functional segregation within these regions on an individual level. Additionally, while most work has focused on a shared semantic network, less has examined shared regions for processing communicative intent. To address these questions, functional and structural MRI data were collected from 24 adult participants while viewing videos of an experimenter producing communicative, Participant-Directed Gestures (PDG) (e.g., "Hello, come here"), noncommunicative Self-adaptor Gestures (SG) (e.g., smoothing hair), and three written text conditions: (1) Participant-Directed Sentences (PDS), matched in content to PDG, (2) Third-person Sentences (3PS), describing a character's actions from a third-person perspective, and (3) meaningless sentences, Jabberwocky (JW). Surface-based conjunction and individual functional region of interest analyses identified shared neural activation between gesture (PDGvsSG) and language processing using two different language contrasts. Conjunction analyses of gesture (PDGvsSG) and Third-person Sentences versus Jabberwocky revealed overlap within left anterior and posterior superior temporal sulcus (STS). Conjunction analyses of gesture and Participant-Directed Sentences to Third-person Sentences revealed regions sensitive to communicative intent, including the left middle and posterior STS and left inferior frontal gyrus. Further, parametric modulation using participants' ratings of stimuli revealed sensitivity of left posterior STS to individual perceptions of communicative intent in gesture. These data highlight an important role of the STS in processing participant-directed communicative intent through gesture and language. Hum Brain Mapp 37:3444-3461, 2016. © 2016 Wiley Periodicals, Inc.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.