Abstract. Classification with rejection is well understood for classifiers which provide explicit class probabilities. The situation is more complicated for popular deterministic classifiers such as learning vector quantisation schemes: albeit reject options using simple distance-based geometric measures were proposed [4], their local scaling behaviour is unclear for complex problems. Here, we propose a local threshold selection strategy which automatically adjusts suitable threshold values for reject options in prototype-based classifiers from given data. We compare this local threshold strategy to a global choice on artificial and benchmark data sets; we show that local thresholds enhance the classification results in comparison to global ones, and they better approximate optimal Bayesian rejection in cases where the latter is available. Keywords: prototype-based reject, classification, local thresholds
MotivationLearning vector quantisation (LVQ) [9] constitutes a powerful and efficient method for multi-class classification tasks which, due to its representation of models in terms of prototypes, is particularly suited for on-line scenarios or lifelong learning [8]. While classical LVQ models have been introduced on heuristic grounds, modern variants are based on cost-function models like generalized LVQ (GLVQ) [12], or robust soft LVQ (RSLVQ) [15] with guarantees on generalization performance and learning dynamics [2,13]. One particular success story links LVQ classifiers to simultaneous metric learners which enrich the classifier with interpretable feature weighting terms or a direct classifier visualisation [13,14]. Still, LVQ classifiers face the problem that real world data do not necessarily allow an unambiguous classification: overlap in the data, outliers, noise, or similar effects can be observed frequently where wrong classifications are unavoidable. A wrong classification can be more costly than postponing a decision and gathering new evidence like in medical diagnostics. Mathematically, such settings can be modelled by introducing a reject option for a classifier: instead of a decision, rejecting is possible for cases with low certainty. This setting has formally been analysed by Chow [3], deriving an optimum decision rule depending on the costs of a reject in comparison to a wrong classification. While this early approach