This paper studies the analytical properties of the reinforcement learning model proposed in Erev and Roth (1998), also termed cumulative reinforcement learning in Laslier et al. (2001). The stochastic model of learning accounts for two main elements: the Law of E¤ect (positive reinforcement of actions that perform well) and the Power Law of Practice (learning curves tend to be steeper initially).The paper establishes a relation between the learning process and the underlying deterministic replicator equation. The main results show that if the solution trajectories of the latter converge su¢ ciently fast, then the probability that all the realizations of the learning process over a given spell of time, possibly in…nite, becomes arbitrarily close to one, from some time on. In particular, the paper shows that the property of fast convergence is always satis…ed in proximity of a strict Nash equilibrium.The results also provide an explicit estimate of the approximation error that could prove to be useful in empirical analysis.JEL: C72, C92, D83.