The progressive ratio (PR) lever-press task serves as a benchmark for assessing goal-oriented motivation. However, a well-recognized limitation of the PR task is that only a single data point, known as the breakpoint, is obtained from an entire session as a barometer of motivation. Because the breakpoint is defined as the final ratio of responses achieved in a PR session, variations in choice behavior during the PR task cannot be captured. We addressed this limitation by constructing four reinforcement learning models: a Simple Q- learning model, an Asymmetric model with two learning rates, a Perseverance model with choice traces, and a Perseverance model without learning. These models incorporated three behavioral choices: reinforced and non-reinforced lever presses and void magazine nosepokes (MNPs), because we noticed that mice performed frequent MNPs during PR tasks. The best model was the Perseverance model, which predicted a gradual reduction in amplitudes of reward prediction errors (RPEs) upon void MNPs. We confirmed the prediction experimentally with fiber photometry of extracellular dopamine (DA) dynamics in the ventral striatum of mice using a fluorescent protein (genetically encoded GPCR activation-based DA sensor: GRABDA2m). We verified application of the model by acute intraperitoneal injection of low-dose methamphetamine (METH) before a PR task, which increased the frequency of MNPs during the PR session without changing the breakpoint. The Perseverance model captured behavioral modulation as a result of increased initial action values, which are customarily set to zero and disregarded in reinforcement learning analysis. Our findings suggest that the Perseverance model reveals effects of psychoactive drugs on choice behaviors during PR tasks.