Abstract. This article looks at the applications of Turing's Legacy in computation, particularly to the theory of algorithmic randomness, where classical mathematical concepts such as measure could be made computational. It also traces Turing's anticipation of this theory in an early manuscript.