Abstract-This paper proposes a tradeoff between computational time, sample complexity, and statistical accuracy that applies to statistical estimators based on convex optimization. When we have a large amount of data, we can exploit excess samples to decrease statistical risk, to decrease computational cost, or to trade off between the two. We propose to achieve this tradeoff by varying the amount of smoothing applied to the optimization problem. This work uses regularized linear regression as a case study to argue for the existence of this tradeoff both theoretically and experimentally. We also apply our method to describe a tradeoff in an image interpolation problem.Index Terms-Smoothing methods, statistical estimation, convex optimization, regularized regression, image interpolation, resource tradeoffs I. MOTIVATION M ASSIVE DATA presents an obvious challenge to statistical algorithms. We expect that the computational effort needed to process a data set increases with its size. The amount of computational power available, however, is growing slowly relative to sample sizes. As a consequence, large-scale problems of practical interest require increasingly more time to solve. This creates a demand for new algorithms that offer better performance when presented with large data sets.While it seems natural that larger problems require more effort to solve, Shalev-Shwartz and Srebro [1] showed that their algorithm for learning a support vector classifier actually becomes faster as the amount of training data increases. This and more recent works support an emerging viewpoint that treats data as a computational resource. That is, we should be able to exploit additional data to improve the performance of statistical algorithms.We consider statistical problems solved through convex optimization and propose the following approach: We can smooth statistical optimization problems more and more aggressively as the amount of available data increases. By controlling the amount of smoothing, we can exploit the additional data to decrease statistical risk, decrease computational cost, or trade off between the two. Our prior work [2] examined a similar time-data tradeoff achieved by applying a dual-smoothing method to (noiseless) regularized linear inverse problems. This paper generalizes those results, allowing for noisy measurements. The result is a tradeoff in computational time, sample size, and statistical accuracy.We use regularized linear regression problems as a specific example to illustrate our principle. We provide theoretical and numerical evidence that supports the existence of a time-data tradeoff achievable through aggressive smoothing of convex optimization problems in the dual domain. Our realization of the tradeoff relies on recent work in convex geometry that allows for precise analysis of statistical risk. In particular, we recognize the work done by Amelunxen et al.[3] to identify phase transitions in regularized linear inverse problems and the extension to noisy problems by Oymak and Hassibi [4]. While we...