“…One such example is Kaplan, Sauer, and Sosa (2017), where natural language methods were used with deep reinforcement learning to play Atari games. In addition to this, recent work on using transformer models with reinforcement learning includes: Parisotto et al (2020), Noever, Ciolino, and Kalin (2020) train GPT-2 (Radford et al 2019) on the PGN format to learn chess, Ciolino, Kalin, and Noever (2020) trained GPT-2 in a similar way to learn Go, and Stein, Filchenkov, and Asadulaev (2020) used Transformers for Deep Q-learning to play Atari games. Krause et al (2020) introduced a coding scheme to improve small language models.…”