We generalise the α-Ramsey cardinals introduced in Holy and Schlicht (2018) for cardinals α to arbitrary ordinals α, and answer several questions posed in that paper. In particular, we show that α-Ramseys are downwards absolute to the core model K for all α of uncountable cofinality, that strategic ω-Ramsey cardinals are equiconsistent with remarkable cardinals and that strategic α-Ramsey cardinals are equiconsistent with measurable cardinals for all α > ω. We also show that the n-Ramseys satisfy indescribability properties and use them to provide a game-theoretic characterisation of completely ineffable cardinals, as well as establishing further connections between the α-Ramsey cardinals and the Ramsey-like cardinals introduced in Gitman (2011), Feng (1990), and Sharpe and Welch (2011).
Monitoring machine learning models once they are deployed is challenging. It is even more challenging to decide when to retrain models in realcase scenarios when labeled data is beyond reach, and monitoring performance metrics becomes unfeasible. In this work, we use non-parametric bootstrapped uncertainty estimates and SHAP values to provide explainable uncertainty estimation as a technique that aims to monitor the deterioration of machine learning models in deployment environments, as well as determine the source of model deterioration when target labels are not available. Classical methods are purely aimed at detecting distribution shift, which can lead to false positives in the sense that the model has not deteriorated despite a shift in the data distribution. To estimate model uncertainty we construct prediction intervals using a novel bootstrap method, which improves upon the work of Kumar & Srivastava (2012). We show that both our model deterioration detection system as well as our uncertainty estimation method achieve better performance than the current state-of-the-art. Finally, we use explainable AI techniques to gain an understanding of the drivers of model deterioration. We release an open source Python package, doubt, which implements our proposed methods, as well as the code used to reproduce our experiments.
Monitoring machine learning models once they are deployed
is challenging. It is even more challenging to decide when
to retrain models in real-case scenarios when labeled data is
beyond reach, and monitoring performance metrics becomes
unfeasible. In this work, we use non-parametric bootstrapped
uncertainty estimates and SHAP values to provide explainable
uncertainty estimation as a technique that aims to monitor
the deterioration of machine learning models in deployment
environments, as well as determine the source of model deteri-
oration when target labels are not available. Classical methods
are purely aimed at detecting distribution shift, which can lead
to false positives in the sense that the model has not deterio-
rated despite a shift in the data distribution. To estimate model
uncertainty we construct prediction intervals using a novel
bootstrap method, which improves previous state-of-the-art
work. We show that both our model deterioration detection
system as well as our uncertainty estimation method achieve
better performance than the current state-of-the-art. Finally,
we use explainable AI techniques to gain an understanding
of the drivers of model deterioration. We release an open
source Python package, doubt, which implements our pro-
posed methods, as well as the code used to reproduce our
experiments.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.