“…Following up on the works of Lai et al and Berry et al, Lai et al gave a review of adaptive enrichment designs, “briefly for clinical trials in new drug development and in more detail for comparativeness effectiveness trials involving approved treatments.” They also used the ideas of Lai and Liao and multiarm bandit theory, developed by Lai and Robbins and Lai, to introduce a new group sequential enrichment design, which uses AR and GLR statistics to “fulfill multiple objectives, which include (i) treating accrued patients with the best (yet unknown) available treatment, (ii) developing a treatment strategy for future patients, and (iii) demonstrating that the strategy developed indeed has better treatment effect than the historical mean effect of SOC plus a predetermined threshold.” They note that, because of the need for informed consent, the clinical trial needs to use randomization in a double blind setting, and the “randomization probability , determined at the l th interim analysis, of assigning a patient in group j to treatment k cannot be too small to suggest obvious inferiority of the treatments being tried, that is for some 0 < ϵ < 1/ K .” Using this constraint, they derive an AR scheme, called ϵ ‐greedy scheme in reinforcement learning, from multiarmed bandit theory. This randomization scheme is easy to implement, in contrast with the Thompson sampling scheme in the Bayesian approach that requires Markov chain Monte Carlo to implement and is also less efficient than ϵ ‐greedy sampling schemes.…”