Learning Combinatorial Optimization Algorithms over Graphs
About
The design of good heuristics or approximation algorithms for NP-hard combinatorial optimization problems often requires significant specialized knowledge and trial-and-error. Can we automate this challenging, tedious process, and learn the algorithms instead? In many real-world applications, it is typically the case that the same optimization problem is solved again and again on a regular basis, maintaining the same problem structure but differing in the data. This provides an opportunity for learning heuristic algorithms that exploit the structure of such recurring problems. In this paper, we propose a unique combination of reinforcement learning and graph embedding to address this challenge. The learned greedy policy behaves like a meta-algorithm that incrementally constructs a solution, and the action is determined by the output of a graph embedding network capturing the current state of the solution. We show that our framework can be applied to a diverse range of optimization problems over graphs, and learns effective algorithms for the Minimum Vertex Cover, Maximum Cut and Traveling Salesman problems.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Traveling Salesman Problem (TSP) | TSP n=100 10K instances (test) | Objective Value8.31 | 52 | |
| Graph Edit Distance | AIDS 20-30 nodes v1 | Objective Error56.5 | 7 | |
| Graph Edit Distance | AIDS 30-50 nodes v1 | Objective Error110 | 7 | |
| Graph Edit Distance | AIDS 50+ nodes v1 | Objective Error183.9 | 7 | |
| Hamiltonian Cycle Problem | FHCP-500/600 (test) | Cycle Found Rate0.00e+0 | 7 | |
| DAG scheduling | TPC-H 50 (test) | Objective Value1.06e+8 | 6 | |
| DAG scheduling | TPC-H 100 (test) | Objective Score1.73e+8 | 6 | |
| DAG scheduling | TPC-H 150 (test) | Objective Score2.48e+8 | 6 | |
| Maximum Independent Set | Cora (test) | MIS Size1.38e+3 | 4 | |
| Maximum Independent Set | Citeseer (test) | Independent Set Size1.71e+3 | 4 |