“…In the context of multi-stream HMM/ANN speech recognition [1,9,10], the logic behind using the entropy in the output vector from each MLP classifier to weight the evidence supplied by that MLP [3,4,5,7,8,11,13] is that uncertain outputs with a flat distribution will have high entropy, while a confident peaked distribution will have low entropy. However, it has been observed that while MLP output entropy does usually increase with noise level, at high noise levels an MLP may also output a high probability that the noisy data comes from just the one or two "attractor" classes which happen to be closest to the noisy input in feature space.…”