2018
DOI: 10.48550/arxiv.1806.00550
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Swiss Army Infinitesimal Jackknife

Abstract: The error or variability of machine learning algorithms is often assessed by repeatedly re-fitting a model with different weighted versions of the observed data. The ubiquitous tools of cross-validation (CV) and the bootstrap are examples of this technique. These methods are powerful in large part due to their model agnosticism but can be slow to run on modern, large data sets due to the need to repeatedly re-fit the model. In this work, we use a linear approximation to the dependence of the fitting procedure … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 11 publications
0
6
0
Order By: Relevance
“…• Scalable Bayesian Inference: Methods inspired from "Infinitesimal Jackknife" (Giordano et al 2019) and "Deep Ensembles" (D'Angelo and Fortuin 2021).…”
Section: Robustness and Reproducibilitymentioning
confidence: 99%
“…• Scalable Bayesian Inference: Methods inspired from "Infinitesimal Jackknife" (Giordano et al 2019) and "Deep Ensembles" (D'Angelo and Fortuin 2021).…”
Section: Robustness and Reproducibilitymentioning
confidence: 99%
“…For example, Hampel et al [12] use influence functions to study the impact of data contamination on statistical procedures. In machine learning, influence functions have been used by [4] to reason about the stability of kernel regression, by [16] to study perturbations of neural networks, and by [11] to obtain inexpensive approximations of resampling procedures like cross-validation and the bootstrap. Ting and Brochu [25] consider single-shot sampling, as compared to our iterative approach, using influence functions, motivating their strategy as minimizing the variance of an asymptotically linear estimator.…”
Section: Related Workmentioning
confidence: 99%
“…Our assumptions are similar to the ones made in [11]. We write the Hessian of the empirical risk minimization problem over the empirical distribution P M at parameter θ as…”
Section: Assumptionsmentioning
confidence: 99%
“…Of course, optimizing over σ (e.g. using cross-validation, jackknife, or their approximate surrogates [54,55,56]) may provide better results. For our method as well as GLK, we do not optimize over the convex combination weights (uniform weights are assigned).…”
Section: Empirical Evaluationsmentioning
confidence: 99%