Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Learning to Solve Constraint Satisfaction Problems with Recurrent Transformer

About

Constraint satisfaction problems (CSPs) are about finding values of variables that satisfy the given constraints. We show that Transformer extended with recurrence is a viable approach to learning to solve CSPs in an end-to-end manner, having clear advantages over state-of-the-art methods such as Graph Neural Networks, SATNet, and some neuro-symbolic models. With the ability of Transformer to handle visual input, the proposed Recurrent Transformer can straightforwardly be applied to visual constraint reasoning problems while successfully addressing the symbol grounding problem. We also show how to leverage deductive knowledge of discrete constraints in the Transformer's inductive learning to achieve sample-efficient learning and semi-supervised learning for CSPs.

Zhun Yang, Adam Ishay, Joohyung Lee• 2023

Related benchmarks

TaskDatasetResultRank
Sudoku SolvingSudoku In-distribution (test)
Accuracy100
6
Sudoku SolvingSudoku OOD - RRN (test)
Instance Solved Rate32.9
6
Showing 2 of 2 rows

Other info

Follow for update