ENIGMA is a learning-based method for guiding given clause selection in saturationbased theorem provers. Clauses from many proof searches are classified as positive and negative based on their participation in the proofs. An efficient classification model is trained on this data, using fast feature-based characterization of the clauses . The learned model is then tightly linked with the core prover and used as a basis of a new parameterized evaluation heuristic that provides fast ranking of all generated clauses. The approach is evaluated on the E prover and the CASC 2016 AIM benchmark, showing a large increase of E's performance.
Introduction: Theorem Proving and LearningState-of-the-art resolution/superposition automated theorem provers (ATPs) such as Vampire [15] and E [20] are today's most advanced tools for general reasoning across a variety of mathematical and scientific domains. The stronger the performance of such tools, the more realistic become tasks such as full computer understanding and automated development of complicated mathematical theories, and verification of software, hardware and engineering designs. While performance of ATPs has steadily grown over the past years due to a number of human-designed improvements, it is still on average far behind the performance of trained mathematicians. Their advanced knowledge-based proof finding is an enigma, which is unlikely to be deciphered and programmed completely manually in near future.On large corpora such as Flyspeck, Mizar and Isabelle, the ATP progress has been mainly due to learning how to select the most relevant knowledge, based on many previous proofs [10,12,1,2]. Learning from many proofs has also recently become a very useful method for automated finding of parameters of ATP strategies [22,9,19,16], and for learning of sequences of tactics in interactive theorem provers (ITPs) [7]. Several experiments with the compact leanCoP [18] system have recently shown that directly using trained machine learner for internal clause selection can significantly prune the search space and solve additional problems [24,11,5]. An obvious next step is to implement efficient learning-based clause selection also inside the strongest superposition-based ATPs.In this work, we introduce ENIGMA -Efficient learNing-based Internal Guidance MAchine for state-of-the-art saturation-based ATPs. The method applies fast machine learning algorithms to a large number of proofs, and uses the trained classifier together with simpler heuristics to evaluate the millions of clauses generated during the resolution/superposition proof search. This way, the theorem prover automatically takes into account thousands of previous successes and failures that it has seen in previous problems, similarly to trained humans. Thanks to a carefully chosen efficient learning/evaluation method and its tight integration with the core ATP (in our case the E prover), the penalty for this ubiquitous knowledge-based internal proof guidance is very low. This in turn very significantly improves the per...