“…However, for general measures µ not necessarily consisting of a finite number of Dirac measures, the literature on the existence of global minima is very limited. There exist positive results for the approximation of functions in the space L p ([0, 1] d , • p ) with shallow feedforward ANNs using heavyside activation [KKV03], the approximation of one-dimensional Lipschitz continuous target functions with shallow feedforward ANNs using ReLU activation and the standard mean square error [JR22], and the approximation of multi-dimensional, continuous target functions with shallow residual ANNs using ReLU activation, see [DJK23]. Conversely, for several common (smooth) activations such as the standard logistic activation, softplus, arctan, hyperbolic tangent and softsign there, generally, do not exist minimizers in the optimization landscape for smooth target functions (or even polynomials), see [PRV21] and [GJL22].…”