“…Strict minimizes the number of large errors [37], in that, for any dag G and data (f , w), if g = Strict(f, w) is an isotonic function on G, then there is a C > 0 such that g has more vertices with regression error ≥ C than does Strict(f, w), and for any D > C, g and Strict(f, w) have the same number of vertices with regression error ≥ D (there might be a d < C where g has fewer vertices with regression error ≥ d than does Strict, but the emphasis is on large errors). For example, for the function with values 3, 1, 2.5 and weights 2, 2, 1, Strict is 2, 2, 2.5, as are all L p isotonic regressions for 1 < p < ∞, but all of the other L ∞ regression mappings considered here have a nonzero error for the third value.…”