This work unifies the analysis of various randomized methods for solving linear and nonlinear in-
verse problems by framing the problem in a stochastic optimization setting. By doing so, we show
that many randomized methods are variants of a sample average approximation. More importantly,
we are able to prove a single theoretical result that guarantees the asymptotic convergence for a
variety of randomized methods. Additionally, viewing randomized methods as a sample average ap-
proximation enables us to prove, for the first time, a single non-asymptotic error result that holds for
randomized methods under consideration. Another important consequence of our unified framework
is that it allows us to discover new randomization methods. We present various numerical results
for linear, nonlinear, algebraic, and PDE-constrained inverse problems that verify the theoretical
convergence results and provide a discussion on the apparently different convergence rates and the
behavior for various randomized methods.