Benchmarks are important for comparing performance of optimisation algorithms, but we can select instances that present our algorithm favourably, and dismiss those on which our algorithm under-performs. Also related are automated design of algorithms, which use problem instances (benchmarks) to train an algorithm: careful choice of instances is needed for the algorithm to generalise. We sweep parameter se ings of di erential evolution to applied to the BBOB benchmarks. Several benchmark functions are highly correlated. is may lead to the false conclusion that an algorithm performs well in general, when it performs poorly on a few key instances. ese correlations vary with the number of evaluations.