“…LSTM generalized well though, requiring only the 30 shortest exemplars (n ≤ 10) of the CSL a n b n c n to correctly predict the possible continuations of sequence prefixes for n up to 1000 and more. A combination of a decoupled extended Kalman filter (Kalman, 1960;Williams, 1992b;Puskorius and Feldkamp, 1994;Feldkamp et al, 1998;Haykin, 2001;Feldkamp et al, 2003) and an LSTM RNN (Pérez-Ortiz et al, 2003) learned to deal correctly with values of n up to 10 million and more. That is, after training the network was able to read sequences of 30,000,000 symbols and more, one symbol at a time, and finally detect the subtle differences between legal strings such as a 10,000,000 b 10,000,000 c 10,000,000 and very similar but illegal strings such as a 10,000,000 b 9,999,999 c 10,000,000 .…”