Premise Selection for Theorem Proving by Deep Graph Embedding
About
We propose a deep learning-based approach to the problem of premise selection: selecting mathematical statements relevant for proving a given conjecture. We represent a higher-order logic formula as a graph that is invariant to variable renaming but still fully preserves syntactic and semantic information. We then embed the graph into a vector via a novel embedding method that preserves the information of edge ordering. Our approach achieves state-of-the-art results on the HolStep dataset, improving the classification accuracy from 83% to 90.3%.
Mingzhe Wang, Yihe Tang, Jian Wang, Jia Deng• 2017
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Premise Selection | HolStep (test) | Accuracy90.3 | 8 | |
| Proof step classification | HolStep (test) | Accuracy90.3 | 5 |
Showing 2 of 2 rows