Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SATformer: Transformer-Based UNSAT Core Learning

About

This paper introduces SATformer, a novel Transformer-based approach for the Boolean Satisfiability (SAT) problem. Rather than solving the problem directly, SATformer approaches the problem from the opposite direction by focusing on unsatisfiability. Specifically, it models clause interactions to identify any unsatisfiable sub-problems. Using a graph neural network, we convert clauses into clause embeddings and employ a hierarchical Transformer-based model to understand clause correlation. SATformer is trained through a multi-task learning approach, using the single-bit satisfiability result and the minimal unsatisfiable core (MUC) for UNSAT problems as clause supervision. As an end-to-end learning-based satisfiability classifier, the performance of SATformer surpasses that of NeuroSAT significantly. Furthermore, we integrate the clause predictions made by SATformer into modern heuristic-based SAT solvers and validate our approach with a logic equivalence checking task. Experimental results show that our SATformer can decrease the runtime of existing solvers by an average of 21.33%.

Zhengyuan Shi, Min Li, Yi Liu, Sadaf Khan, Junhua Huang, Hui-Ling Zhen, Mingxuan Yuan, Qiang Xu (1) __INSTITUTION_8__ The Chinese University of Hong Kong, (2) Huawei Noah's Ark Lab)• 2022

Related benchmarks

TaskDatasetResultRank
SAT solvingJNH structured SAT family
MRPP r-tilde1.75
15
SAT solvingPARITY structured SAT family
MRPP r-tilde0.73
15
SAT solvingrandom 3-SAT 50
MRPP r~0.88
12
SAT solvingrandom 3-SAT 31–60
MRPP r~84
12
SAT solvingrandom 3-SAT 5–15
MRPP r~100
12
SAT solvingrandom 3-SAT (16–30)
MRPP r~0.89
12
SAT solvingrandom 3-SAT 61–100
MRPP r~0.83
12
SAT solvingrandom 3-SAT 100
MRPP r~0.82
12
SAT solvingPRET
MRPP r~1
9
SAT solvingPHOLE
MRPP r˜1
9
Showing 10 of 14 rows

Other info

Follow for update