In changing environments animals must adaptively select actions to achieve their goals. In tasks involving goal-directed action selection, striatal neural activity has been shown to represent the value of competing actions. Striatal representations of action value could potentially bias responses toward actions of higher value. However, no study to date has demonstrated the direct impact of distinct striatal pathways in goal-directed action selection. Here we show in mice that transient optogenetic stimulation of dorsal striatal dopamine D1 and D2 receptor-expressing neurons during decision-making introduces opposing biases in the distribution of choices. The effect of stimulation on choice is dependent on recent reward history and mimics an additive change in the action value. While stimulation prior to and during movement initiation produces a robust bias in choice behavior, this bias is significantly diminished when stimulation is delayed after response initiation. Together, our data demonstrate the role of striatal activity in goal-directed action selection.
Although systems involved in attentional selection have been studied extensively, much less is known about non-selective systems. To study these preparatory mechanisms, we compared activity in auditory cortex elicited by sounds while rats performed an auditory task (“engaged”) with activity elicited by identical stimuli while subjects were awake but not performing a task (“passive”). Surprisingly, we found that engagement suppressed responses, an effect opposite in sign to that elicited by selective attention. In the auditory thalamus, however, engagement enhanced spontaneous firing rates but did not affect evoked responses. These results demonstrate that in auditory cortex, neural activity cannot be viewed simply as a limited resource allocated in greater measure as the state of the animal passes from somnolent to passively listening to engaged and attentive. Instead the engaged condition possesses a characteristic and distinct neural signature in which sound-evoked responses are paradoxically suppressed.
The basal ganglia, a group of subcortical nuclei, play a crucial role in decision making by selecting actions and evaluating their outcomes1,2. While much is known about the function of the basal ganglia circuitry in selection1,3,4, how these nuclei contribute to outcome evaluation is less clear. Here we show that neurons in the habenula-projecting globus pallidus (GPh) are essential for evaluating action outcomes and are regulated by a specific set of inputs from the basal ganglia. We found in a classical conditioning task that individual mouse GPh neurons bidirectionally encode whether an outcome is better or worse than expected. Mimicking these evaluation signals with optogenetic inhibition or excitation is sufficient to reinforce or discourage actions in a decision making task. Moreover, cell-type-specific synaptic manipulations revealed that the inhibitory and excitatory inputs to the GPh are necessary for mice to appropriately evaluate positive and negative feedback, respectively. Finally, using rabies virus-assisted monosynaptic tracing5, we discovered that the GPh is embedded in a basal ganglia circuit wherein it receives inhibitory input from both striosomal and matrix compartments of the striatum, and excitatory input from the “limbic” regions of the subthalamic nucleus (STN). Our results provide the first direct evidence that information about the selection and evaluation of actions is channelled through distinct sets of basal ganglia circuits, with the GPh representing a key locus where information of opposing valence is integrated to determine whether action outcomes are better or worse than expected.
Rules encompass cue-action-outcome associations used to guide decisions and strategies in a specific context. Subregions of the frontal cortex including the orbitofrontal cortex (OFC) and dorsomedial prefrontal cortex (dmPFC) are implicated in rule learning, although changes in structural connectivity underlying rule learning are poorly understood. We imaged OFC axonal projections to dmPFC during training in a multiple choice foraging task and used a reinforcement learning model to quantify explore–exploit strategy use and prediction error magnitude. Here we show that rule training, but not experience of reward alone, enhances OFC bouton plasticity. Baseline bouton density and gains during training correlate with rule exploitation, while bouton loss correlates with exploration and scales with the magnitude of experienced prediction errors. We conclude that rule learning sculpts frontal cortex interconnectivity and adjusts a thermostat for the explore–exploit balance.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.