At initialization, artificial neural networks (ANNs) are equivalent to Gaussian processes in the infinite-width limit (14; 11), thus connecting them to kernel methods. We prove that the evolution of an ANN during training can also be described by a kernel: during gradient descent on the parameters of an ANN, the network function f θ (which maps input vectors to output vectors) follows the kernel gradient of the functional cost (which is convex, in contrast to the parameter cost) w.r.t. a new kernel: the Neural Tangent Kernel (NTK). This kernel is central to describe the generalization features of ANNs. While the NTK is random at initialization and varies during training, in the infinite-width limit it converges to an explicit limiting kernel and it stays constant during training. This makes it possible to study the training of ANNs in function space instead of parameter space. Convergence of the training can then be related to the positive-definiteness of the limiting NTK. We prove the positive-definiteness of the limiting NTK when the data is supported on the sphere and the non-linearity is non-polynomial. We then focus on the setting of least-squares regression and show that in the infinitewidth limit, the network function f θ follows a linear differential equation during training. The convergence is fastest along the largest kernel principal components of the input data with respect to the NTK, hence suggesting a theoretical motivation for early stopping. Finally we study the NTK numerically, observe its behavior for wide networks, and compare it to the infinite-width limit.
We rigorously prove the existence and the conformal invariance of scaling limits of the magnetization and multi-point spin correlations in the critical Ising model on arbitrary simply connected planar domains. This solves a number of conjectures coming from the physical and the mathematical literature. The proof relies on convergence results for discrete holomorphic spinor observables and probabilistic techniques.
We study the critical Ising model on the square lattice in bounded simply connected domains with + and free boundary conditions. We relate the energy density of the model to a discrete fermionic spinor and compute its scaling limit by discrete complex analysis methods. As a consequence, we obtain a simple exact formula for the scaling limit of the energy field onepoint function in terms of the hyperbolic metric. This confirms the predictions originating in physics, but also provides a higher precision.
Abstract. We show how to combine our earlier results to deduce strong convergence of the interfaces in the planar critical Ising model and its random-cluster representation to Schramm's SLE curves with parameter κ = 3 and κ = 16/3 respectively.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.