SATformer: Transformer-Based UNSAT Core Learning
About
This paper introduces SATformer, a novel Transformer-based approach for the Boolean Satisfiability (SAT) problem. Rather than solving the problem directly, SATformer approaches the problem from the opposite direction by focusing on unsatisfiability. Specifically, it models clause interactions to identify any unsatisfiable sub-problems. Using a graph neural network, we convert clauses into clause embeddings and employ a hierarchical Transformer-based model to understand clause correlation. SATformer is trained through a multi-task learning approach, using the single-bit satisfiability result and the minimal unsatisfiable core (MUC) for UNSAT problems as clause supervision. As an end-to-end learning-based satisfiability classifier, the performance of SATformer surpasses that of NeuroSAT significantly. Furthermore, we integrate the clause predictions made by SATformer into modern heuristic-based SAT solvers and validate our approach with a logic equivalence checking task. Experimental results show that our SATformer can decrease the runtime of existing solvers by an average of 21.33%.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| SAT solving | JNH structured SAT family | MRPP r-tilde1.75 | 15 | |
| SAT solving | PARITY structured SAT family | MRPP r-tilde0.73 | 15 | |
| SAT solving | random 3-SAT 50 | MRPP r~0.88 | 12 | |
| SAT solving | random 3-SAT 31–60 | MRPP r~84 | 12 | |
| SAT solving | random 3-SAT 5–15 | MRPP r~100 | 12 | |
| SAT solving | random 3-SAT (16–30) | MRPP r~0.89 | 12 | |
| SAT solving | random 3-SAT 61–100 | MRPP r~0.83 | 12 | |
| SAT solving | random 3-SAT 100 | MRPP r~0.82 | 12 | |
| SAT solving | PRET | MRPP r~1 | 9 | |
| SAT solving | PHOLE | MRPP r˜1 | 9 |