“…On one hand, many approaches utilize human-selected features and expert-designed strategies, including Situation Based Strategic Positioning (Reis et al, 2001 ), multi-agent positioning mechanism (Akiyama and Noda, 2008 ), coordination system based on setplays (Mota and Reis, 2007 ), positioning based on Delaunay Triangulation (Akiyama and Noda, 2007 ), and Voronoi diagrams (Prokopenko and Wang, 2017 ). Others involve well-optimized defense and attack behaviors in popular code bases such as Agent2d (Akiyama and Nakashima, 2013 ) and Gliders2d (Prokopenko and Wang, 2019a , b ). On the other hand, machine learning approaches have been applied in RCSS environment as well, e.g., a reinforcement learning approach (Riedmiller et al, 2001 , 2008 ; Gabel et al, 2009 ), online planning with tree search method (Akiyama et al, 2012 ), and MAXQ value function decomposition for online planning (Bai et al, 2015 ).…”