Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Neural Set Function Extensions: Learning with Discrete Functions in High Dimensions

About

Integrating functions on discrete domains into neural networks is key to developing their capability to reason about discrete objects. But, discrete domains are (1) not naturally amenable to gradient-based optimization, and (2) incompatible with deep learning architectures that rely on representations in high-dimensional vector spaces. In this work, we address both difficulties for set functions, which capture many important discrete problems. First, we develop a framework for extending set functions onto low-dimensional continuous domains, where many extensions are naturally defined. Our framework subsumes many well-known extensions as special cases. Second, to avoid undesirable low-dimensional neural network bottlenecks, we convert low-dimensional extensions into representations in high-dimensional spaces, taking inspiration from the success of semidefinite programs for combinatorial optimization. Empirically, we observe benefits of our extensions for unsupervised neural combinatorial optimization, in particular with high-dimensional representations.

Nikolaos Karalias, Joshua Robinson, Andreas Loukas, Stefanie Jegelka• 2022

Related benchmarks

TaskDatasetResultRank
Minimum Vertex CoverCOLLAB (test)
AR*1.008
16
Minimum Vertex CoverIMDB-BINARY (test)
AR*1
12
Minimum Vertex CoverTwitter (test)
AR*1.028
12
Maximum CutBA 200-300 graphs (test)
MCut Value693.5
11
Maximum Independent SetCOLLAB first 1000 graphs (test)
Approximation Ratio1
10
Maximum Clique ProblemENZYMES (test)
AR*0.933
10
Maximum Clique ProblemIMDB-BINARY (test)
AR*0.961
10
Maximum Independent SetENZYMES 0.3 (test)
AR*0.821
10
Maximum Independent SetPROTEINS (test)
AR*0.903
10
Maximum Independent SetIMDB-BINARY (test)
AR*0.917
10
Showing 10 of 13 rows

Other info

Follow for update