“…Accordingly, understanding SGD's behavior has been crucial to its reliable application in machine learning and beyond. As a result, SGD's theory has greatly advanced from a variety of perspectives: global convergence analysis Gower et al, 2020;Khaled & Richtárik, 2020;Mertikopoulos et al, 2020;Patel, 2020), local convergence analysis (Mertikopoulos et al, 2020), greedy and global complexity analysis (Gower et al, 2020;Khaled & Richtárik, 2020), asymptotic weak convergence (Wang et al, 2021), and saddle point analysis (Fang et al, 2019;Mertikopoulos et al, 2020;Jin et al, 2021). While all of these perspectives add new dimensions to our understanding of SGD, the global convergence analysis of SGD is the foundation as it dictates whether local analyses, complexity analyses or saddle point analyses are even warranted.…”